Switch to: References

Add citations

You must login to add citations.
  1. The Oxford Handbook of Causal Reasoning.Michael Waldmann (ed.) - 2017 - Oxford, England: Oxford University Press.
    Causal reasoning is one of our most central cognitive competencies, enabling us to adapt to our world. Causal knowledge allows us to predict future events, or diagnose the causes of observed facts. We plan actions and solve problems using knowledge about cause-effect relations. Without our ability to discover and empirically test causal theories, we would not have made progress in various empirical sciences. In the past decades, the important role of causal knowledge has been discovered in many areas of cognitive (...)
    Download  
     
    Export citation  
     
    Bookmark   8 citations  
  • Emotion and decision-making: affect-driven belief systems in anxiety and depression.Martin P. Paulus & Angela J. Yu - 2012 - Trends in Cognitive Sciences 16 (9):476-483.
    Download  
     
    Export citation  
     
    Bookmark   11 citations  
  • Artificial intelligence and modern planned economies: a discussion on methods and institutions.Spyridon Samothrakis - forthcoming - AI and Society:1-12.
    Interest in computerised central economic planning (CCEP) has seen a resurgence, as there is strong demand for an alternative vision to modern free (or not so free) market liberal capitalism. Given the close links of CCEP with what we would now broadly call artificial intelligence (AI)—e.g. optimisation, game theory, function approximation, machine learning, automated reasoning—it is reasonable to draw direct analogues and perform an analysis that would help identify what commodities and institutions we should see for a CCEP programme to (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • Resource Rationality.Thomas F. Icard - manuscript
    Theories of rational decision making often abstract away from computational and other resource limitations faced by real agents. An alternative approach known as resource rationality puts such matters front and center, grounding choice and decision in the rational use of finite resources. Anticipated by earlier work in economics and in computer science, this approach has recently seen rapid development and application in the cognitive sciences. Here, the theory of rationality plays a dual role, both as a framework for normative assessment (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • Knowledge representation and acquisition for ethical AI: challenges and opportunities.Vaishak Belle - 2023 - Ethics and Information Technology 25 (1):1-12.
    Machine learning (ML) techniques have become pervasive across a range of different applications, and are now widely used in areas as disparate as recidivism prediction, consumer credit-risk analysis, and insurance pricing. Likewise, in the physical world, ML models are critical components in autonomous agents such as robotic surgeons and self-driving cars. Among the many ethical dimensions that arise in the use of ML technology in such applications, analyzing morally permissible actions is both immediate and profound. For example, there is the (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • Artificial virtuous agents in a multi-agent tragedy of the commons.Jakob Stenseke - 2022 - AI and Society:1-18.
    Although virtue ethics has repeatedly been proposed as a suitable framework for the development of artificial moral agents, it has been proven difficult to approach from a computational perspective. In this work, we present the first technical implementation of artificial virtuous agents in moral simulations. First, we review previous conceptual and technical work in artificial virtue ethics and describe a functionalistic path to AVAs based on dispositional virtues, bottom-up learning, and top-down eudaimonic reward. We then provide the details of a (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • Reward tampering problems and solutions in reinforcement learning: a causal influence diagram perspective.Tom Everitt, Marcus Hutter, Ramana Kumar & Victoria Krakovna - 2021 - Synthese 198 (Suppl 27):6435-6467.
    Can humans get arbitrarily capable reinforcement learning agents to do their bidding? Or will sufficiently capable RL agents always find ways to bypass their intended objectives by shortcutting their reward signal? This question impacts how far RL can be scaled, and whether alternative paradigms must be developed in order to build safe artificial general intelligence. In this paper, we study when an RL agent has an instrumental goal to tamper with its reward process, and describe design principles that prevent instrumental (...)
    Download  
     
    Export citation  
     
    Bookmark   4 citations  
  • Decentralized MDPs with sparse interactions.Francisco S. Melo & Manuela Veloso - 2011 - Artificial Intelligence 175 (11):1757-1789.
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • Minimax real-time heuristic search.Sven Koenig - 2001 - Artificial Intelligence 129 (1-2):165-197.
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • Faster Teaching via POMDP Planning.Anna N. Rafferty, Emma Brunskill, Thomas L. Griffiths & Patrick Shafto - 2016 - Cognitive Science 40 (6):1290-1332.
    Human and automated tutors attempt to choose pedagogical activities that will maximize student learning, informed by their estimates of the student's current knowledge. There has been substantial research on tracking and modeling student learning, but significantly less attention on how to plan teaching actions and how the assumed student model impacts the resulting plans. We frame the problem of optimally selecting teaching actions using a decision-theoretic approach and show how to formulate teaching as a partially observable Markov decision process planning (...)
    Download  
     
    Export citation  
     
    Bookmark   3 citations  
  • The Nature of Belief-Directed Exploratory Choice in Human Decision-Making.W. Bradley Knox, A. Ross Otto, Peter Stone & Bradley C. Love - 2011 - Frontiers in Psychology 2.
    Download  
     
    Export citation  
     
    Bookmark   9 citations  
  • “Guess what I'm doing”: Extending legibility to sequential decision tasks.Miguel Faria, Francisco S. Melo & Ana Paiva - 2024 - Artificial Intelligence 330 (C):104107.
    Download  
     
    Export citation  
     
    Bookmark  
  • Simplified Risk-aware Decision Making with Belief-dependent Rewards in Partially Observable Domains.Andrey Zhitnikov & Vadim Indelman - 2022 - Artificial Intelligence 312 (C):103775.
    Download  
     
    Export citation  
     
    Bookmark  
  • Representations for robot knowledge in the KnowRob framework.Moritz Tenorth & Michael Beetz - 2017 - Artificial Intelligence 247 (C):151-169.
    Download  
     
    Export citation  
     
    Bookmark   4 citations  
  • An affective mobile robot educator with a full-time job.Illah R. Nourbakhsh, Judith Bobenage, Sebastien Grange, Ron Lutz, Roland Meyer & Alvaro Soto - 1999 - Artificial Intelligence 114 (1-2):95-124.
    Download  
     
    Export citation  
     
    Bookmark  
  • Optimal Behavior is Easier to Learn than the Truth.Ronald Ortner - 2016 - Minds and Machines 26 (3):243-252.
    We consider a reinforcement learning setting where the learner is given a set of possible models containing the true model. While there are algorithms that are able to successfully learn optimal behavior in this setting, they do so without trying to identify the underlying true model. Indeed, we show that there are cases in which the attempt to find the true model is doomed to failure.
    Download  
     
    Export citation  
     
    Bookmark  
  • Finding the optimal exploration-exploitation trade-off online through Bayesian risk estimation and minimization.Stewart Jamieson, Jonathan P. How & Yogesh Girdhar - 2024 - Artificial Intelligence 330 (C):104096.
    Download  
     
    Export citation  
     
    Bookmark  
  • Autonomous agents modelling other agents: A comprehensive survey and open problems.Stefano V. Albrecht & Peter Stone - 2018 - Artificial Intelligence 258 (C):66-95.
    Download  
     
    Export citation  
     
    Bookmark   15 citations  
  • Conformant plans and beyond: Principles and complexity.Blai Bonet - 2010 - Artificial Intelligence 174 (3-4):245-269.
    Download  
     
    Export citation  
     
    Bookmark   4 citations  
  • Action understanding as inverse planning.Chris L. Baker, Rebecca Saxe & Joshua B. Tenenbaum - 2009 - Cognition 113 (3):329-349.
    Download  
     
    Export citation  
     
    Bookmark   90 citations  
  • Shared Representations as Coordination Tools for Interaction.Giovanni Pezzulo - 2011 - Review of Philosophy and Psychology 2 (2):303-333.
    Why is interaction so simple? This article presents a theory of interaction based on the use of shared representations as “coordination tools” (e.g., roundabouts that facilitate coordination of drivers). By aligning their representations (intentionally or unintentionally), interacting agents help one another to solve interaction problems in that they remain predictable, and offer cues for action selection and goal monitoring. We illustrate how this strategy works in a joint task (building together a tower of bricks) and discuss its requirements from a (...)
    Download  
     
    Export citation  
     
    Bookmark   11 citations  
  • A survey of inverse reinforcement learning: Challenges, methods and progress.Saurabh Arora & Prashant Doshi - 2021 - Artificial Intelligence 297 (C):103500.
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  • Recursively modeling other agents for decision making: A research perspective.Prashant Doshi, Piotr Gmytrasiewicz & Edmund Durfee - 2020 - Artificial Intelligence 279 (C):103202.
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • Inferring the intentional states of autonomous virtual agents.Peter C. Pantelis, Chris L. Baker, Steven A. Cholewiak, Kevin Sanik, Ari Weinstein, Chia-Chien Wu, Joshua B. Tenenbaum & Jacob Feldman - 2014 - Cognition 130 (3):360-379.
    Download  
     
    Export citation  
     
    Bookmark   5 citations  
  • Probabilistic reasoning about epistemic action narratives.Fabio Aurelio D'Asaro, Antonis Bikakis, Luke Dickens & Rob Miller - 2020 - Artificial Intelligence 287 (C):103352.
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • Partially observable Markov decision processes with imprecise parameters.Hideaki Itoh & Kiyohiko Nakamura - 2007 - Artificial Intelligence 171 (8-9):453-490.
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  • Risk-aware shielding of Partially Observable Monte Carlo Planning policies.Giulio Mazzi, Alberto Castellini & Alessandro Farinelli - 2023 - Artificial Intelligence 324 (C):103987.
    Download  
     
    Export citation  
     
    Bookmark  
  • Approximability and efficient algorithms for constrained fixed-horizon POMDPs with durative actions.Majid Khonji - 2023 - Artificial Intelligence 323 (C):103968.
    Download  
     
    Export citation  
     
    Bookmark  
  • Expanding horizons in reinforcement learning for curious exploration and creative planning.Dale Zhou & Aaron M. Bornstein - 2024 - Behavioral and Brain Sciences 47:e118.
    Curiosity and creativity are expressions of the trade-off between leveraging that with which we are familiar or seeking out novelty. Through the computational lens of reinforcement learning, we describe how formulating the value of information seeking and generation via their complementary effects on planning horizons formally captures a range of solutions to striking this balance.
    Download  
     
    Export citation  
     
    Bookmark  
  • Rationalizing predictions by adversarial information calibration.Lei Sha, Oana-Maria Camburu & Thomas Lukasiewicz - 2023 - Artificial Intelligence 315 (C):103828.
    Download  
     
    Export citation  
     
    Bookmark  
  • A dynamic epistemic framework for reasoning about conformant probabilistic plans.Yanjun Li, Barteld Kooi & Yanjing Wang - 2019 - Artificial Intelligence 268 (C):54-84.
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  • Robotic manipulation of multiple objects as a POMDP.Joni Pajarinen & Ville Kyrki - 2017 - Artificial Intelligence 247:213-228.
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • An integrated approach to solving influence diagrams and finite-horizon partially observable decision processes.Eric A. Hansen - 2021 - Artificial Intelligence 294 (C):103431.
    Download  
     
    Export citation  
     
    Bookmark  
  • POMDP-based control of workflows for crowdsourcing.Peng Dai, Christopher H. Lin, Mausam & Daniel S. Weld - 2013 - Artificial Intelligence 202 (C):52-85.
    Download  
     
    Export citation  
     
    Bookmark  
  • Contingent planning under uncertainty via stochastic satisfiability.Stephen M. Majercik & Michael L. Littman - 2003 - Artificial Intelligence 147 (1-2):119-162.
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • Knowledge-based programs as succinct policies for partially observable domains.Bruno Zanuttini, Jérôme Lang, Abdallah Saffidine & François Schwarzentruber - 2020 - Artificial Intelligence 288 (C):103365.
    Download  
     
    Export citation  
     
    Bookmark   3 citations  
  • Computing rank dependent utility in graphical models for sequential decision problems.Gildas Jeantet & Olivier Spanjaard - 2011 - Artificial Intelligence 175 (7-8):1366-1389.
    Download  
     
    Export citation  
     
    Bookmark  
  • Analyzing generalized planning under nondeterminism.Vaishak Belle - 2022 - Artificial Intelligence 307 (C):103696.
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  • Epistemic uncertainty aware semantic localization and mapping for inference and belief space planning.Vladimir Tchuiev & Vadim Indelman - 2023 - Artificial Intelligence 319 (C):103903.
    Download  
     
    Export citation  
     
    Bookmark  
  • Regression and progression in stochastic domains.Vaishak Belle & Hector J. Levesque - 2020 - Artificial Intelligence 281 (C):103247.
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  • Reasoning about uncertain parameters and agent behaviors through encoded experiences and belief planning.Akinobu Hayashi, Dirk Ruiken, Tadaaki Hasegawa & Christian Goerick - 2020 - Artificial Intelligence 280 (C):103228.
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • A synthesis of automated planning and reinforcement learning for efficient, robust decision-making.Matteo Leonetti, Luca Iocchi & Peter Stone - 2016 - Artificial Intelligence 241 (C):103-130.
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • Exploiting symmetries for single- and multi-agent Partially Observable Stochastic Domains.Byung Kon Kang & Kee-Eung Kim - 2012 - Artificial Intelligence 182-183 (C):32-57.
    Download  
     
    Export citation  
     
    Bookmark  
  • From Knowledge-based Programs to Graded Belief-based Programs, Part I: On-line Reasoning.Noël Laverny & Jérôme Lang - 2005 - Synthese 147 (2):277-321.
    Knowledge-based programs (KBPs) are a powerful notion for expressing action policies in which branching conditions refer to implicit knowledge and call for a deliberation task at execution time. However, branching conditions in KBPs cannot refer to possibly erroneous beliefs or to graded belief, such as “if my belief that φ holds is high then do some action α else perform some sensing action β”.
    Download  
     
    Export citation  
     
    Bookmark   4 citations  
  • Common Bayesian Models for Common Cognitive Issues.Francis Colas, Julien Diard & Pierre Bessière - 2010 - Acta Biotheoretica 58 (2-3):191-216.
    How can an incomplete and uncertain model of the environment be used to perceive, infer, decide and act efficiently? This is the challenge that both living and artificial cognitive systems have to face. Symbolic logic is, by its nature, unable to deal with this question. The subjectivist approach to probability is an extension to logic that is designed specifically to face this challenge. In this paper, we review a number of frequently encountered cognitive issues and cast them into a common (...)
    Download  
     
    Export citation  
     
    Bookmark   7 citations  
  • Affect control processes: Intelligent affective interaction using a partially observable Markov decision process.Jesse Hoey, Tobias Schröder & Areej Alhothali - 2016 - Artificial Intelligence 230 (C):134-172.
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • Geometric backtracking for combined task and motion planning in robotic systems.Julien Bidot, Lars Karlsson, Fabien Lagriffoul & Alessandro Saffiotti - 2017 - Artificial Intelligence 247 (C):229-265.
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  • A technical survey on statistical modelling and design methods for crowdsourcing quality control.Yuan Jin, Mark Carman, Ye Zhu & Yong Xiang - 2020 - Artificial Intelligence 287 (C):103351.
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • Planning to see: A hierarchical approach to planning visual actions on a robot using POMDPs.Mohan Sridharan, Jeremy Wyatt & Richard Dearden - 2010 - Artificial Intelligence 174 (11):704-725.
    Download  
     
    Export citation  
     
    Bookmark  
  • A conflict-directed approach to chance-constrained mixed logical linear programming.Cheng Fang & Brian C. Williams - 2023 - Artificial Intelligence 323 (C):103972.
    Download  
     
    Export citation  
     
    Bookmark