Results for 'Alignment'

983 found
Order:
  1. AI Alignment vs. AI Ethical Treatment: Ten Challenges.Adam Bradley & Bradford Saad - manuscript
    A morally acceptable course of AI development should avoid two dangers: creating unaligned AI systems that pose a threat to humanity and mistreating AI systems that merit moral consideration in their own right. This paper argues these two dangers interact and that if we create AI systems that merit moral consideration, simultaneously avoiding both of these dangers would be extremely challenging. While our argument is straightforward and supported by a wide range of pretheoretical moral judgments, it has far-reaching moral implications (...)
    Download  
     
    Export citation  
     
    Bookmark   3 citations  
  2. Alignment and commitment in joint action.Matthew Rachar - 2018 - Philosophical Psychology 31 (6):831-849.
    Important work on alignment systems has been applied to philosophical work on joint action by Tollefsen and Dale. This paper builds from and expands on their work. The first aim of the paper is to spell out how the empirical research on alignment may be integrated into philosophical theories of joint action. The second aim is then to develop a successful characterization of joint action, which spells out the difference between genuine joint action and simpler forms of coordination (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  3. Is Alignment Unsafe?Cameron Domenico Kirk-Giannini - 2024 - Philosophy and Technology 37 (110):1–4.
    Inchul Yum (2024) argues that the widespread adoption of language agent architectures would likely increase the risk posed by AI by simplifying the process of aligning artificial systems with human values and thereby making it easier for malicious actors to use them to cause a variety of harms. Yum takes this to be an example of a broader phenomenon: progress on the alignment problem is likely to be net safety-negative because it makes artificial systems easier for malicious actors to (...)
    Download  
     
    Export citation  
     
    Bookmark  
  4.  79
    Expanding AI and AI Alignment Discourse: An Opportunity for Greater Epistemic Inclusion.A. E. Williams - manuscript
    The AI and AI alignment communities have been instrumental in addressing existential risks, developing alignment methodologies, and promoting rationalist problem-solving approaches. However, as AI research ventures into increasingly uncertain domains, there is a risk of premature epistemic convergence, where prevailing methodologies influence not only the evaluation of ideas but also determine which ideas are considered within the discourse. This paper examines critical epistemic blind spots in AI alignment research, particularly the lack of predictive frameworks to differentiate problems (...)
    Download  
     
    Export citation  
     
    Bookmark  
  5. Aligning with the Good.Benjamin Mitchell-Yellin - 2015 - Journal of Ethics and Social Philosophy (2):1-8.
    IN “CONSTRUCTIVISM, AGENCY, AND THE PROBLEM of Alignment,” Michael Bratman considers how lessons from the philosophy of action bear on the question of how best to construe the agent’s standpoint in the context of a constructivist theory of practical reasons. His focus is “the problem of alignment”: “whether the pressures from the general constructivism will align with the pressures from the theory of agency” (Bratman 2012: 81). He thus brings two lively literatures into dialogue with each other. This (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  6.  50
    We Should Not Align Quantitative Measures with Stakeholder Values.Miguel Ohnesorge - forthcoming - Philosophy of Science.
    There is a growing consensus among philosophers that quantifying value-laden concepts can be epistemically successful and politically legitimate if all value-laden choices in the process of quantification are aligned with stakeholder values. I argue that proponents of this view have failed to argue for its basic premise: successful quantification is sufficiently unconstrained so that it can be achieved along multiple stakeholder-specific pathways. I then challenge this premise by considering a rare example of successful value-laden quantification in seismology. Seismologists quantified earthquake (...)
    Download  
     
    Export citation  
     
    Bookmark  
  7. Variable Value Alignment by Design; averting risks with robot religion.Jeffrey White - forthcoming - Embodied Intelligence 2023.
    Abstract: One approach to alignment with human values in AI and robotics is to engineer artiTicial systems isomorphic with human beings. The idea is that robots so designed may autonomously align with human values through similar developmental processes, to realize project ideal conditions through iterative interaction with social and object environments just as humans do, such as are expressed in narratives and life stories. One persistent problem with human value orientation is that different human beings champion different values as (...)
    Download  
     
    Export citation  
     
    Bookmark  
  8. Mindshaping, Coordination, and Intuitive Alignment.Daniel I. Perez-Zapata & Ian A. Apperly - forthcoming - In Tad Zawidzki, Routledge Handbook of Mindshaping.
    In this chapter, we will summarize recent empirical results highlighting how different groups of people solve pure coordination games. Such games are traditionally studied in behavioural economics, where two people need to coordinate without communicating with each other. Our results suggest that coordination choices vary across groups of people, and that people can adapt flexibly to these differences in order to coordinate between groups. We propose that pure coordination games are a useful empirical platform for studying aspects of mindshaping. Drawing (...)
    Download  
     
    Export citation  
     
    Bookmark  
  9. AI, alignment, and the categorical imperative.Fritz McDonald - 2023 - AI and Ethics 3:337-344.
    Tae Wan Kim, John Hooker, and Thomas Donaldson make an attempt, in recent articles, to solve the alignment problem. As they define the alignment problem, it is the issue of how to give AI systems moral intelligence. They contend that one might program machines with a version of Kantian ethics cast in deontic modal logic. On their view, machines can be aligned with human values if such machines obey principles of universalization and autonomy, as well as a deontic (...)
    Download  
     
    Export citation  
     
    Bookmark  
  10. Discovering Our Blind Spots and Cognitive Biases in AI Research and Alignment.A. E. Williams - manuscript
    The challenge of AI alignment is not just a technological issue but fundamentally an epistemic one. AI safety research predominantly relies on empirical validation, often detecting failures only after they manifest. However, certain risks—such as deceptive alignment and goal misspecification—may not be empirically testable until it is too late, necessitating a shift toward leading-indicator logical reasoning. This paper explores how mainstream AI research systematically filters out deep epistemic insight, hindering progress in AI safety. We assess the rarity of (...)
    Download  
     
    Export citation  
     
    Bookmark  
  11. Aligning Patient’s Ideas of a Good Life with Medically Indicated Therapies in Geriatric Rehabilitation Using Smart Sensors.Cristian Timmermann, Frank Ursin, Christopher Predel & Florian Steger - 2021 - Sensors 21 (24):8479.
    New technologies such as smart sensors improve rehabilitation processes and thereby increase older adults’ capabilities to participate in social life, leading to direct physical and mental health benefits. Wearable smart sensors for home use have the additional advantage of monitoring day-to-day activities and thereby identifying rehabilitation progress and needs. However, identifying and selecting rehabilitation priorities is ethically challenging because physicians, therapists, and caregivers may impose their own personal values leading to paternalism. Therefore, we develop a discussion template consisting of a (...)
    Download  
     
    Export citation  
     
    Bookmark  
  12. The elusive transformation of research and innovation. The overlooked complexities of value alignment and joint responsibility.Giovanni De Grandis - 2025 - In Giovanni De Grandis & Anne Blanchard, The Fragility of Responsibility. Norway’s Transformative Agenda for Research, Innovation and Business. Berlin, Boston: De Gruyter. pp. 83-116.
    RRI is a broad concept that is subject to different interpretations. This chapter focuses on the view of RRI as a transformative ideal for reforming the research and innovation system in the service of public interest. This is the normatively strong view of RRI that has attracted many policy-makers and young researchers but left cold many senior researchers and innovators. The transformative vision of RRI has failed to materialise, and RRI remains a marginal reality, even in Norway, where arguably the (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  13.  49
    Aligning AI with the Universal Formula for Balanced Decision-Making.Angelito Malicse - manuscript
    -/- Aligning AI with the Universal Formula for Balanced Decision-Making -/- Introduction -/- Artificial Intelligence (AI) represents a highly advanced form of automated information processing, capable of analyzing vast amounts of data, identifying patterns, and making predictive decisions. However, the effectiveness of AI depends entirely on the integrity of its inputs, processing mechanisms, and decision-making frameworks. If AI is programmed without a foundational understanding of natural laws, it risks reinforcing misinformation, bias, and societal imbalance. -/- Angelito Malicse’s universal formula, particularly (...)
    Download  
     
    Export citation  
     
    Bookmark  
  14. Improve Alignment of Research Policy and Societal Values.Peter Novitzky, Michael J. Bernstein, Vincent Blok, Robert Braun, Tung Tung Chan, Wout Lamers, Anne Loeber, Ingeborg Meijer, Ralf Lindner & Erich Griessler - 2020 - Science 369 (6499):39-41.
    Historically, scientific and engineering expertise has been key in shaping research and innovation policies, with benefits presumed to accrue to society more broadly over time. But there is persistent and growing concern about whether and how ethical and societal values are integrated into R&I policies and governance, as we confront public disbelief in science and political suspicion toward evidence-based policy-making. Erosion of such a social contract with science limits the ability of democratic societies to deal with challenges presented by new, (...)
    Download  
     
    Export citation  
     
    Bookmark   5 citations  
  15. Values in science and AI alignment research.Leonard Dung - manuscript
    Roughly, empirical AI alignment research (AIA) is an area of AI research which investigates empirically how to design AI systems in line with human goals. This paper examines the role of non-epistemic values in AIA. It argues that: (1) Sciences differ in the degree to which values influence them. (2) AIA is strongly value-laden. (3) This influence of values is managed inappropriately and thus threatens AIA’s epistemic integrity and ethical beneficence. (4) AIA should strive to achieve value transparency, critical (...)
    Download  
     
    Export citation  
     
    Bookmark  
  16. AI Alignment Problem: “Human Values” don’t Actually Exist.Alexey Turchin - manuscript
    Abstract. The main current approach to the AI safety is AI alignment, that is, the creation of AI whose preferences are aligned with “human values.” Many AI safety researchers agree that the idea of “human values” as a constant, ordered sets of preferences is at least incomplete. However, the idea that “humans have values” underlies a lot of thinking in the field; it appears again and again, sometimes popping up as an uncritically accepted truth. Thus, it deserves a thorough (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  17. The linguistic dead zone of value-aligned agency, natural and artificial.Travis LaCroix - 2024 - Philosophical Studies:1-23.
    The value alignment problem for artificial intelligence (AI) asks how we can ensure that the “values”—i.e., objective functions—of artificial systems are aligned with the values of humanity. In this paper, I argue that linguistic communication is a necessary condition for robust value alignment. I discuss the consequences that the truth of this claim would have for research programmes that attempt to ensure value alignment for AI systems—or, more loftily, those programmes that seek to design robustly beneficial or (...)
    Download  
     
    Export citation  
     
    Bookmark  
  18. Disagreement, AI alignment, and bargaining.Harry R. Lloyd - forthcoming - Philosophical Studies:1-31.
    New AI technologies have the potential to cause unintended harms in diverse domains including warfare, judicial sentencing, biomedicine and governance. One strategy for realising the benefits of AI whilst avoiding its potential dangers is to ensure that new AIs are properly ‘aligned’ with some form of ‘alignment target.’ One danger of this strategy is that – dependent on the alignment target chosen – our AIs might optimise for objectives that reflect the values only of a certain subset of (...)
    Download  
     
    Export citation  
     
    Bookmark  
  19.  21
    The End of Resonance: A Structural Critique of AI Alignment and the Imminent Collapse of Human Judgement.Jinho Kim - manuscript
    This paper introduces a novel critique of the AI alignment problem, grounded in structural judgemental philosophy. While traditional AI alignment frameworks assume that aligning machine behavior with human goals is sufficient, we argue that this view omits the deeper structure of human judgement itself—namely, the triadic architecture of affectivity, constructibility, and resonance. As Large Language Models (LLMs) evolve without consciousness yet continue to simulate judgement, they threaten to displace the very structures that make human judgement possible. We warn (...)
    Download  
     
    Export citation  
     
    Bookmark  
  20. Democratic education: Aligning curriculum, pedagogy, assessment and school governance.Gilbert Burgh - 2003 - In Philip Cam, Philosophy, democracy and education. pp. 101–120.
    Matthew Lipman claims that the community of inquiry is an exemplar of democracy in action. To many proponents the community of inquiry is considered invaluable for achieving desirable social and political ends through education for democracy. But what sort of democracy should we be educating for? In this paper I outline three models of democracy: the liberal model, which emphasises rights and duties, and draws upon pre-political assumptions about freedom; communitarianism, which focuses on identity and participation in the creation of (...)
    Download  
     
    Export citation  
     
    Bookmark   6 citations  
  21. Biomedical ontology alignment: An approach based on representation learning.Prodromos Kolyvakis, Alexandros Kalousis, Barry Smith & Dimitris Kiritsis - 2018 - Journal of Biomedical Semantics 9 (21).
    While representation learning techniques have shown great promise in application to a number of different NLP tasks, they have had little impact on the problem of ontology matching. Unlike past work that has focused on feature engineering, we present a novel representation learning approach that is tailored to the ontology matching task. Our approach is based on embedding ontological terms in a high-dimensional Euclidean space. This embedding is derived on the basis of a novel phrase retrofitting strategy through which semantic (...)
    Download  
     
    Export citation  
     
    Bookmark  
  22. (1 other version)An Enactive Approach to Value Alignment in Artificial Intelligence: A Matter of Relevance.Michael Cannon - 2021 - In Vincent C. Müller, Philosophy and Theory of AI. Springer Cham. pp. 119-135.
    The “Value Alignment Problem” is the challenge of how to align the values of artificial intelligence with human values, whatever they may be, such that AI does not pose a risk to the existence of humans. Existing approaches appear to conceive of the problem as "how do we ensure that AI solves the problem in the right way", in order to avoid the possibility of AI turning humans into paperclips in order to “make more paperclips” or eradicating the human (...)
    Download  
     
    Export citation  
     
    Bookmark  
  23. In Conversation with Artificial Intelligence: Aligning language Models with Human Values.Atoosa Kasirzadeh - 2023 - Philosophy and Technology 36 (2):1-24.
    Large-scale language technologies are increasingly used in various forms of communication with humans across different contexts. One particular use case for these technologies is conversational agents, which output natural language text in response to prompts and queries. This mode of engagement raises a number of social and ethical questions. For example, what does it mean to align conversational agents with human norms or values? Which norms or values should they be aligned with? And how can this be accomplished? In this (...)
    Download  
     
    Export citation  
     
    Bookmark   20 citations  
  24. The Role of Foundational Relations in the Alignment of Biomedical Ontologies.Barry Smith & Cornelius Rosse - 2004 - In Stefan Schulze-Kremer, MedInfo. IOS Press. pp. 444-448.
    The Foundational Model of Anatomy (FMA) symbolically represents the structural organization of the human body from the macromolecular to the macroscopic levels, with the goal of providing a robust and consistent scheme for classifying anatomical entities that is designed to serve as a reference ontology in biomedical informatics. Here we articulate the need for formally clarifying the is-a and part-of relations in the FMA and similar ontology and terminology systems. We diagnose certain characteristic errors in the treatment of these relations (...)
    Download  
     
    Export citation  
     
    Bookmark   30 citations  
  25. Robustness to Fundamental Uncertainty in AGI Alignment.G. G. Worley Iii - 2020 - Journal of Consciousness Studies 27 (1-2):225-241.
    The AGI alignment problem has a bimodal distribution of outcomes with most outcomes clustering around the poles of total success and existential, catastrophic failure. Consequently, attempts to solve AGI alignment should, all else equal, prefer false negatives (ignoring research programs that would have been successful) to false positives (pursuing research programs that will unexpectedly fail). Thus, we propose adopting a policy of responding to points of philosophical and practical uncertainty associated with the alignment problem by limiting and (...)
    Download  
     
    Export citation  
     
    Bookmark  
  26. Engineering the trust machine. Aligning the concept of trust in the context of blockchain applications.Eva Pöll - 2024 - Ethics and Information Technology 26 (2):1-16.
    Complex technology has become an essential aspect of everyday life. We rely on technology as part of basic infrastructure and repeatedly for tasks throughout the day. Yet, in many cases the relation surpasses mere reliance and evolves to trust in technology. A new, disruptive technology is blockchain. It claims to introduce trustless relationships among its users, aiming to eliminate the need for trust altogether—even being described as “the trust machine”. This paper presents a proposal to adjust the concept of trust (...)
    Download  
     
    Export citation  
     
    Bookmark  
  27.  27
    Reprogramming Society: Aligning Human Learning, Education, and AI with the Universal Law of Balance.Angelito Malicse - manuscript
    -/- Reprogramming Society: Aligning Human Learning, Education, and AI with the Universal Law of Balance -/- Introduction -/- Throughout history, human societies have struggled with misinformation, irrational decision-making, and social imbalance. The root cause of these issues lies in the way human minds are programmed from birth. Negative thinking and behavior are not inherent traits but the result of flawed learning systems that fail to align with the universal law of balance in nature. To correct this, a holistic transformation of (...)
    Download  
     
    Export citation  
     
    Bookmark  
  28.  21
    Aligning Tariff Policy with the Universal Formula for Free Will.Angelito Malicse - manuscript
    Download  
     
    Export citation  
     
    Bookmark  
  29. Stop re-inventing the wheel: or how ELSA and RRI can align.Mark Ryan & Vincent Blok - 2023 - Journal of Responsible Innovation (x):x.
    Ethical, Legal and Social Aspects (ELSA) originated in the 4thEuropean Research Framework Programme (1994) andresponsible research and innovation (RRI) from the EC researchagenda in 2010. ELSA has received renewed attention inEuropean funding schemes and research. This raises the questionof how these two approaches to social responsibility relate toone another and if there is the possibility to align. There is aneed to evaluate the relationship/overlap between ELSA and RRIbecause there is a possibility that new ELSA research will reinventthe wheel if it (...)
    Download  
     
    Export citation  
     
    Bookmark   3 citations  
  30. Robustness to fundamental uncertainty in AGI alignment.I. I. I. G. Gordon Worley - manuscript
    The AGI alignment problem has a bimodal distribution of outcomes with most outcomes clustering around the poles of total success and existential, catastrophic failure. Consequently, attempts to solve AGI alignment should, all else equal, prefer false negatives (ignoring research programs that would have been successful) to false positives (pursuing research programs that will unexpectedly fail). Thus, we propose adopting a policy of responding to points of metaphysical and practical uncertainty associated with the alignment problem by limiting and (...)
    Download  
     
    Export citation  
     
    Bookmark  
  31. A pragmatic approach to scientific change: transfer, alignment, influence.Stefano Canali - 2022 - European Journal for Philosophy of Science 12 (3):1-25.
    I propose an approach that expands philosophical views of scientific change, on the basis of an analysis of contemporary biomedical research and recent developments in the philosophy of scientific change. Focusing on the establishment of the exposome in epidemiology as a case study and the role of data as a context for contrasting views on change, I discuss change at conceptual, methodological, material, and social levels of biomedical epistemology. Available models of change provide key resources to discuss this type of (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  32.  22
    Bridging Priorities: Aligning Community Actions with Biodiversity Conservation Goals.Hùng Hòe - 2025 - Xomchim.Com.
    Effective biodiversity conservation increasingly depends on community engagement [2,3]. However, identifying which actions should be prioritized remains a complex challenge. Dean et al. [4] addressed this gap by developing an innovative framework that integrates expert evaluations of ecological impact with community perceptions of feasibility and willingness to act.
    Download  
     
    Export citation  
     
    Bookmark  
  33.  34
    Forestry 5.0: Aligning Human Expertise and AI for Sustainable Forest Futures.Kiến Lửa - 2025 - Xomchim.Com.
    As technological advancement reshapes industries, the forestry sector is poised for a pivotal transformation. Forestry 5.0, building upon the principles of Industry 5.0, introduces a paradigm where human-centered artificial intelligence (AI) collaborates with, rather than replaces, human expertise to promote sustainable, efficient, and resilient forest management [2].
    Download  
     
    Export citation  
     
    Bookmark  
  34. Intelligent AI Personal Trainer: Advanced Posture Alignment & Exercise Companion.A. T. Vivin & Thanabal M. S. - 2025 - Journal of Science Technology and Research (JSTAR) 6 (1):1-15.
    AI Personal Trainer: Advanced Posture Alignment & Exercise Companion is a smart fitness system designed to improve workout accuracy and reduce injury risk through continuous posture monitoring and real-time corrective feedback. Leveraging advanced deep learning models—specifically Recurrent Neural Networks (RNNs) and Long Short-Term Memory (LSTM) networks—the system accurately identifies and evaluates exercises such as squats, curls, flassioni, and planks, ensuring users maintain optimal form. Through computer vision techniques and pose estimation using the Mediapipe framework, skeletal keypoints are extracted from (...)
    Download  
     
    Export citation  
     
    Bookmark  
  35.  61
    The Eden Framework: Exploring Divergence, Alignment, and the Ethical Flow of Information.Tim Grooms - manuscript
    Abstract This paper examines the Eden narrative as an allegory for the interplay of free will, ethical alignment, and the emergence of “dark information.” It argues that God’s will can be understood as divine information—a foundational structure that ensures harmony when adhered to. Through divergence from this information, entropy is introduced, necessitating cyclical renewal. By exploring theological, philosophical, and informational perspectives, this paper highlights the relevance of these concepts in addressing modern challenges, offering actionable frameworks for realignment with foundational (...)
    Download  
     
    Export citation  
     
    Bookmark  
  36. Control and Flexibility of Interactive Alignment: Mobius Syndrome as a Case Study.John Michael, Kathleen Bogart, Kristian Tylen, Joel Krueger, Morten Bech, John R. Ostergaard & Riccardo Fusaroli - 2014 - Cognitive Processing 15 (1):S125-126.
    Download  
     
    Export citation  
     
    Bookmark  
  37. The genetic technologies questionnaire: lay judgments about genetic technologies align with ethical theory, are coherent, and predict behaviour.Svenja Küchenhoff, Johannes Doerflinger & Nora Heinzelmann - 2022 - BMC Medical Ethics 23 (54):1-14.
    -/- Policy regulations of ethically controversial genetic technologies should, on the one hand, be based on ethical principles. On the other hand, they should be socially acceptable to ensure implementation. In addition, they should align with ethical theory. Yet to date we lack a reliable and valid scale to measure the relevant ethical judgements in laypeople. We target this lacuna. -/- We developed a scale based on ethical principles to elicit lay judgments: the Genetic Technologies Questionnaire (GTQ). In two pilot (...)
    Download  
     
    Export citation  
     
    Bookmark  
  38. The Prospect of a Humanitarian Artificial Intelligence: Agency and Value Alignment.Montemayor Carlos - 2023
    In this open access book, Carlos Montemayor illuminates the development of artificial intelligence (AI) by examining our drive to live a dignified life. -/- He uses the notions of agency and attention to consider our pursuit of what is important. His method shows how the best way to guarantee value alignment between humans and potentially intelligent machines is through attention routines that satisfy similar needs. Setting out a theoretical framework for AI Montemayor acknowledges its legal, moral, and political implications (...)
    Download  
     
    Export citation  
     
    Bookmark  
  39. The marriage of astrology and AI: A model of alignment with human values and intentions.Kenneth McRitchie - 2024 - Correlation 36 (1):43-49.
    Astrology research has been using artificial intelligence (AI) to improve the understanding of astrological properties and processes. Like the large language models of AI, astrology is also a language model with a similar underlying linguistic structure but with a distinctive layer of lifestyle contexts. Recent research in semantic proximities and planetary dominance models have helped to quantify effective astrological information. As AI learning and intelligence grows, a major concern is with maintaining its alignment with human values and intentions. Astrology (...)
    Download  
     
    Export citation  
     
    Bookmark  
  40.  36
    The Evolution of Human Consciousness and Alignment with the Cosmic Force.Angelito Malicse - manuscript
    The Evolution of Human Consciousness and Alignment with the Cosmic Force -/- Throughout history, humanity has sought to understand the forces that shape its existence. While many perspectives frame evolution as a purely biological process, an alternative view suggests that human evolution is actively guided by an impersonal cosmic force. This force subtly directs our intellectual, spiritual, and technological advancements, leading us toward a higher state of being. This essay explores the idea that humanity is gradually aligning with this (...)
    Download  
     
    Export citation  
     
    Bookmark  
  41.  18
    Restructuring Profit-Oriented Social Media to Align with the Three Universal Laws of Nature.Angelito Malicse - manuscript
    -/- Restructuring Profit-Oriented Social Media to Align with the Three Universal Laws of Nature -/- By Angelito Malicse -/- In today’s world, social media platforms have become powerful tools for shaping thought, behavior, and society itself. However, the dominance of profit-oriented models in these platforms has led to widespread harm—ranging from misinformation and emotional manipulation to mental health crises and societal division. When assessed through the lens of my Three Universal Laws of Nature—the Law of Karma, the Law of Feedback, (...)
    Download  
     
    Export citation  
     
    Bookmark  
  42. Cross‐cultural variation and perspectivalism: Alignment of two red herrings?Jincai Li - 2023 - Mind and Language 38 (4):1157-1163.
    In this brief reply I respond to criticisms of my book, The referential mechanism of proper names, from Michael Devitt and Nicolo D'Agruma. I focus on the question of whether the perspectivism advocated in the book explains the empirical results there detailed.
    Download  
     
    Export citation  
     
    Bookmark  
  43. Soldierly Virtue: An argument for the restructuring of Western military ethics to align with Aristotelian Virtue Ethics.John Baldari - 2018 - Dissertation, University of Leeds
    Because wars are fought by human beings and not merely machines, a strong virtue ethic is an essential prerequisite for those engaged in combat. From a philosophical perspective, war has historically been seen as separate and outside of the commonly accepted forms of morality. Yet there remains a general, though not well-thought out, sense that those human beings who fight wars should act ethically. Since warfighters are often called upon to contemplate and complete tasks during war that are not normally (...)
    Download  
     
    Export citation  
     
    Bookmark  
  44. Beyond Competence: Why AI Needs Purpose, Not Just Programming.Georgy Iashvili - manuscript
    The alignment problem in artificial intelligence (AI) is a critical challenge that extends beyond the need to align future superintelligent systems with human values. This paper argues that even "merely intelligent" AI systems, built on current-gen technologies, pose existential risks due to their competence-without-comprehension nature. Current AI models, despite their advanced capabilities, lack intrinsic moral reasoning and are prone to catastrophic misalignment when faced with ethical dilemmas, as illustrated by recent controversies. Solutions such as hard-coded censorship and rule-based restrictions (...)
    Download  
     
    Export citation  
     
    Bookmark  
  45. Deontology and Safe Artificial Intelligence.William D’Alessandro - forthcoming - Philosophical Studies:1-24.
    The field of AI safety aims to prevent increasingly capable artificially intelligent systems from causing humans harm. Research on moral alignment is widely thought to offer a promising safety strategy: if we can equip AI systems with appropriate ethical rules, according to this line of thought, they'll be unlikely to disempower, destroy or otherwise seriously harm us. Deontological morality looks like a particularly attractive candidate for an alignment target, given its popularity, relative technical tractability and commitment to harm-avoidance (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  46. Taking Into Account Sentient Non-Humans in AI Ambitious Value Learning: Sentientist Coherent Extrapolated Volition.Adrià Moret - 2023 - Journal of Artificial Intelligence and Consciousness 10 (02):309-334.
    Ambitious value learning proposals to solve the AI alignment problem and avoid catastrophic outcomes from a possible future misaligned artificial superintelligence (such as Coherent Extrapolated Volition [CEV]) have focused on ensuring that an artificial superintelligence (ASI) would try to do what humans would want it to do. However, present and future sentient non-humans, such as non-human animals and possible future digital minds could also be affected by the ASI’s behaviour in morally relevant ways. This paper puts forward Sentientist Coherent (...)
    Download  
     
    Export citation  
     
    Bookmark  
  47. Explicability as an AI Principle: Technology and Ethics in Cooperation.Moto Kamiura - forthcoming - Proceedings of the 39Th Annual Conference of the Japanese Society for Artificial Intelligence, 2025.
    This paper categorizes current approaches to AI ethics into four perspectives and briefly summarizes them: (1) Case studies and technical trend surveys, (2) AI governance, (3) Technologies for AI alignment, (4) Philosophy. In the second half, we focus on the fourth perspective, the philosophical approach, within the context of applied ethics. In particular, the explicability of AI may be an area in which scientists, engineers, and AI developers are expected to engage more actively relative to other ethical issues in (...)
    Download  
     
    Export citation  
     
    Bookmark  
  48. AI Survival Stories: a Taxonomic Analysis of AI Existential Risk.Herman Cappelen, Simon Goldstein & John Hawthorne - forthcoming - Philosophy of Ai.
    Since the release of ChatGPT, there has been a lot of debate about whether AI systems pose an existential risk to humanity. This paper develops a general framework for thinking about the existential risk of AI systems. We analyze a two-premise argument that AI systems pose a threat to humanity. Premise one: AI systems will become extremely powerful. Premise two: if AI systems become extremely powerful, they will destroy humanity. We use these two premises to construct a taxonomy of ‘survival (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  49. ChatGPT: towards AI subjectivity.Kristian D’Amato - 2024 - AI and Society 39:1-15.
    Motivated by the question of responsible AI and value alignment, I seek to offer a uniquely Foucauldian reconstruction of the problem as the emergence of an ethical subject in a disciplinary setting. This reconstruction contrasts with the strictly human-oriented programme typical to current scholarship that often views technology in instrumental terms. With this in mind, I problematise the concept of a technological subjectivity through an exploration of various aspects of ChatGPT in light of Foucault’s work, arguing that current systems (...)
    Download  
     
    Export citation  
     
    Bookmark   3 citations  
  50. Security practices in AI development.Petr Spelda & Vit Stritecky - forthcoming - AI and Society.
    What makes safety claims about general purpose AI systems such as large language models trustworthy? We show that rather than the capabilities of security tools such as alignment and red teaming procedures, it is security practices based on these tools that contributed to reconfiguring the image of AI safety and made the claims acceptable. After showing what causes the gap between the capabilities of security tools and the desired safety guarantees, we critically investigate how AI security practices attempt to (...)
    Download  
     
    Export citation  
     
    Bookmark  
1 — 50 / 983