Switch to: Citations

Add references

You must login to add references.
  1. Utilitarianism.R. M. Hare - 1963 - In Richard Mervyn Hare (ed.), Freedom and reason. Oxford,: Clarendon Press.
    Through consideration of another practical case, this chapter opens the way to a generalization of the method of argument outlined previously. Multilateral cases raise the question of how the interests of all parties can be resolved into a determinate moral conclusion, which brings the discussion to a standpoint that has affinities with classical utilitarianism. Like the principle of universalizability, the form of the utilitarian principle espoused is purely logical. In both cases, the moral substance comes from fleshing out the parties’ (...)
    Download  
     
    Export citation  
     
    Bookmark   31 citations  
  • Understanding Artificial Agency.Leonard Dung - forthcoming - Philosophical Quarterly.
    Which artificial intelligence (AI) systems are agents? To answer this question, I propose a multidimensional account of agency. According to this account, a system's agency profile is jointly determined by its level of goal-directedness and autonomy as well as is abilities for directly impacting the surrounding world, long-term planning and acting for reasons. Rooted in extant theories of agency, this account enables fine-grained, nuanced comparative characterizations of artificial agency. I show that this account has multiple important virtues and is more (...)
    Download  
     
    Export citation  
     
    Bookmark   3 citations  
  • The argument for near-term human disempowerment through AI.Leonard Dung - 2024 - AI and Society:1-14.
    Many researchers and intellectuals warn about extreme risks from artificial intelligence. However, these warnings typically came without systematic arguments in support. This paper provides an argument that AI will lead to the permanent disempowerment of humanity, e.g. human extinction, by 2100. It rests on four substantive premises which it motivates and defends: first, the speed of advances in AI capability, as well as the capability level current systems have already reached, suggest that it is practically possible to build AI systems (...)
    Download  
     
    Export citation  
     
    Bookmark   4 citations  
  • Current cases of AI misalignment and their implications for future risks.Leonard Dung - 2023 - Synthese 202 (5):1-23.
    How can one build AI systems such that they pursue the goals their designers want them to pursue? This is the alignment problem. Numerous authors have raised concerns that, as research advances and systems become more powerful over time, misalignment might lead to catastrophic outcomes, perhaps even to the extinction or permanent disempowerment of humanity. In this paper, I analyze the severity of this risk based on current instances of misalignment. More specifically, I argue that contemporary large language models and (...)
    Download  
     
    Export citation  
     
    Bookmark   5 citations  
  • Machine Learning, Functions and Goals.Patrick Butlin - 2022 - Croatian Journal of Philosophy 22 (66):351-370.
    Machine learning researchers distinguish between reinforcement learning and supervised learning and refer to reinforcement learning systems as “agents”. This paper vindicates the claim that systems trained by reinforcement learning are agents while those trained by supervised learning are not. Systems of both kinds satisfy Dretske’s criteria for agency, because they both learn to produce outputs selectively in response to inputs. However, reinforcement learning is sensitive to the instrumental value of outputs, giving rise to systems which exploit the effects of outputs (...)
    Download  
     
    Export citation  
     
    Bookmark   4 citations  
  • Black Boxes or Unflattering Mirrors? Comparative Bias in the Science of Machine Behaviour.Cameron Buckner - 2023 - British Journal for the Philosophy of Science 74 (3):681-712.
    The last 5 years have seen a series of remarkable achievements in deep-neural-network-based artificial intelligence research, and some modellers have argued that their performance compares favourably to human cognition. Critics, however, have argued that processing in deep neural networks is unlike human cognition for four reasons: they are (i) data-hungry, (ii) brittle, and (iii) inscrutable black boxes that merely (iv) reward-hack rather than learn real solutions to problems. This article rebuts these criticisms by exposing comparative bias within them, in the (...)
    Download  
     
    Export citation  
     
    Bookmark   7 citations  
  • Motivational Internalism and the Challenge of Amoralism.Danielle Bromwich - 2016 - European Journal of Philosophy 24 (2):452-471.
    Motivational internalism is the thesis that captures the commonplace thought that moral judgements are necessarily motivationally efficacious. But this thesis appears to be in tension with another aspect of our ordinary moral experience. Proponents of the contrast thesis, motivational externalism, cite everyday examples of amoralism to demonstrate that it is conceptually possible to be completely unmoved by what seem to be sincere first-person moral judgements. This paper argues that the challenge of amoralism gives us no reason to reject or modify (...)
    Download  
     
    Export citation  
     
    Bookmark   14 citations  
  • Philosophers on Philosophy: The 2020 PhilPapers Survey.David Bourget & David J. Chalmers - 2023 - Philosophers' Imprint 23 (11).
    What are the philosophical views of professional philosophers, and how do these views change over time? The 2020 PhilPapers Survey surveyed around 2000 philosophers on 100 philosophical questions. The results provide a snapshot of the state of some central debates in philosophy, reveal correlations and demographic effects involving philosophers' views, and reveal some changes in philosophers' views over the last decade.
    Download  
     
    Export citation  
     
    Bookmark   42 citations  
  • The Superintelligent Will: Motivation and Instrumental Rationality in Advanced Artificial Agents. [REVIEW]Nick Bostrom - 2012 - Minds and Machines 22 (2):71-85.
    This paper discusses the relation between intelligence and motivation in artificial agents, developing and briefly arguing for two theses. The first, the orthogonality thesis, holds (with some caveats) that intelligence and final goals (purposes) are orthogonal axes along which possible artificial intellects can freely vary—more or less any level of intelligence could be combined with more or less any final goal. The second, the instrumental convergence thesis, holds that as long as they possess a sufficient level of intelligence, agents having (...)
    Download  
     
    Export citation  
     
    Bookmark   44 citations  
  • Recent Work on Motivational Internalism.Fredrik Björklund, Gunnar Björnsson, John Eriksson, Ragnar Francén Olinder & Caj Strandberg - 2012 - Analysis 72 (1):124-137.
    Reviews work on moral judgment motivational internalism from the last two decades.
    Download  
     
    Export citation  
     
    Bookmark   47 citations  
  • Existential risk from AI and orthogonality: Can we have it both ways?Vincent C. Müller & Michael Cannon - 2021 - Ratio 35 (1):25-36.
    The standard argument to the conclusion that artificial intelligence (AI) constitutes an existential risk for the human species uses two premises: (1) AI may reach superintelligent levels, at which point we humans lose control (the ‘singularity claim’); (2) Any level of intelligence can go along with any goal (the ‘orthogonality thesis’). We find that the singularity claim requires a notion of ‘general intelligence’, while the orthogonality thesis requires a notion of ‘instrumental intelligence’. If this interpretation is correct, they cannot be (...)
    Download  
     
    Export citation  
     
    Bookmark   7 citations  
  • An AGI Modifying Its Utility Function in Violation of the Strong Orthogonality Thesis.James D. Miller, Roman Yampolskiy & Olle Häggström - 2020 - Philosophies 5 (4):40.
    An artificial general intelligence (AGI) might have an instrumental drive to modify its utility function to improve its ability to cooperate, bargain, promise, threaten, and resist and engage in blackmail. Such an AGI would necessarily have a utility function that was at least partially observable and that was influenced by how other agents chose to interact with it. This instrumental drive would conflict with the strong orthogonality thesis since the modifications would be influenced by the AGI’s intelligence. AGIs in highly (...)
    Download  
     
    Export citation  
     
    Bookmark   7 citations