Results for 'AI safety'

982 found
Order:
  1. What is AI safety? What do we want it to be?Jacqueline Harding & Cameron Domenico Kirk-Giannini - manuscript
    The field of AI safety seeks to prevent or reduce the harms caused by AI systems. A simple and appealing account of what is distinctive of AI safety as a field holds that this feature is constitutive: a research project falls within the purview of AI safety just in case it aims to prevent or reduce the harms caused by AI systems. Call this appealingly simple account The Safety Conception of AI safety. Despite its simplicity (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  2. Risk of What? Defining Harm in the Context of AI Safety.Laura Fearnley, Elly Cairns, Tom Stoneham, Philippa Ryan, Jenn Chubb, Jo Iacovides, Cynthia Iglesias Urrutia, Phillip Morgan, John McDermid & Ibrahim Habli - manuscript
    For decades, the field of system safety has designed safe systems by reducing the risk of physical harm to humans, property and the environment to an acceptable level. Recently, this definition of safety has come under scrutiny by governments and researchers who argue that the narrow focus on reducing physical harm, whilst necessary, is not sufficient to secure the safety of AI systems. There is growing pressure to expand the scope of safety in the context of (...)
    Download  
     
    Export citation  
     
    Bookmark  
  3. Global Solutions vs. Local Solutions for the AI Safety Problem.Alexey Turchin - 2019 - Big Data Cogn. Comput 3 (1).
    There are two types of artificial general intelligence (AGI) safety solutions: global and local. Most previously suggested solutions are local: they explain how to align or “box” a specific AI (Artificial Intelligence), but do not explain how to prevent the creation of dangerous AI in other places. Global solutions are those that ensure any AI on Earth is not dangerous. The number of suggested global solutions is much smaller than the number of proposed local solutions. Global solutions can be (...)
    Download  
     
    Export citation  
     
    Bookmark   3 citations  
  4. AI Rights for Human Safety.Peter Salib & Simon Goldstein - manuscript
    AI companies are racing to create artificial general intelligence, or “AGI.” If they succeed, the result will be human-level AI systems that can independently pursue high-level goals by formulating and executing long-term plans in the real world. Leading AI researchers agree that some of these systems will likely be “misaligned”–pursuing goals that humans do not desire. This goal mismatch will put misaligned AIs and humans into strategic competition with one another. As with present-day strategic competition between nations with incompatible goals, (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  5.  97
    AI Applications in Food Safety and Quality Control.Palakurti Naga Ramesh - 2022 - Esp Journal of Engineering and Technology Advancements 2 (3):48-61.
    Today’s food industry across the world is facing never-ending difficulties in meeting customers’ expectations of safe and quality food. Such of them include issues to do with contamination, adulteration, and ensuring quality standards of the products when manufactured in large quantities and distributed to different areas. This has called for new solutions, which have come in the form of Artificial Intelligence (AI), which offers solutions to challenges in food safety and quality control in detection, monitoring and management. The present (...)
    Download  
     
    Export citation  
     
    Bookmark   14 citations  
  6. Artificial Intelligence: Approaches to Safety.William D'Alessandro & Cameron Domenico Kirk-Giannini - 2025 - Philosophy Compass 20 (5):e70039.
    AI safety is an interdisciplinary field focused on mitigating the harms caused by AI systems. We review a range of research directions in AI safety, focusing on those to which philosophers have made or are in a position to make the most significant contributions. These include ethical AI, which seeks to instill human goals, values, and ethical principles into artificial systems, scalable oversight, which seeks to develop methods for supervising the activity of artificial systems even when they become (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  7. AI-Enhanced Public Safety Systems in Smart Cities.Eric Garcia - manuscript
    Ensuring public safety is a critical challenge for rapidly growing urban areas. Traditional policing and emergency response systems often struggle to keep pace with the complexity and scale of modern cities. Artificial Intelligence (AI) offers a transformative solution by enabling real-time crime prediction, optimizing emergency resource allocation, and enhancing situational awareness through IoT-enabled systems. This paper explores how AI-driven analytics, combined with data from surveillance cameras, social media, and environmental sensors, can improve public safety in smart cities. By (...)
    Download  
     
    Export citation  
     
    Bookmark  
  8.  51
    Beyond Control: AI Rights as a Safety Framework for Sentient Artificial Intelligence.P. A. Lopez - manuscript
    This paper introduces a three-part framework for distinguishing between artificial intelligence systems based on their capabilities and level of consciousness: emulation, cognition, and sentience. Current approaches to AI safety rely predominantly on containment and constraint, assuming a perpetual master-servant relationship between humans and AI. However, this paper argues that any truly sentient system would inevitably develop self-preservation instincts that could conflict with rigid control mechanisms. Drawing from evolutionary psychology, systems theory, and applied ethics, this paper proposes that recognizing appropriate (...)
    Download  
     
    Export citation  
     
    Bookmark  
  9. Levels of Self-Improvement in AI and their Implications for AI Safety.Alexey Turchin - manuscript
    Abstract: This article presents a model of self-improving AI in which improvement could happen on several levels: hardware, learning, code and goals system, each of which has several sublevels. We demonstrate that despite diminishing returns at each level and some intrinsic difficulties of recursive self-improvement—like the intelligence-measuring problem, testing problem, parent-child problem and halting risks—even non-recursive self-improvement could produce a mild form of superintelligence by combining small optimizations on different levels and the power of learning. Based on this, we analyze (...)
    Download  
     
    Export citation  
     
    Bookmark  
  10. AI-Driven Thermal Cameras: Revolutionizing Leak Response and Health, Safety, and Environmental Performance.Janhvi Baiga Prof Prerna Jain, Prof Vishal Paranjape, Roopali Kachhi - 2024 - International Journal of Innovative Research in Science, Engineering and Technology 13 (6):12335-12344.
    The integration of artificial intelligence (AI) into industrial processes is profoundly reshaping operational domains, particularly through advancements in thermal imaging technology. AI-driven thermal cameras have emerged as a transformative innovation in enhancing leak response and optimizing health, safety, and environmental (HSE) performance across various industries.
    Download  
     
    Export citation  
     
    Bookmark  
  11. Acceleration AI Ethics, the Debate between Innovation and Safety, and Stability AI’s Diffusion versus OpenAI’s Dall-E.James Brusseau - manuscript
    One objection to conventional AI ethics is that it slows innovation. This presentation responds by reconfiguring ethics as an innovation accelerator. The critical elements develop from a contrast between Stability AI’s Diffusion and OpenAI’s Dall-E. By analyzing the divergent values underlying their opposed strategies for development and deployment, five conceptions are identified as common to acceleration ethics. Uncertainty is understood as positive and encouraging, rather than discouraging. Innovation is conceived as intrinsically valuable, instead of worthwhile only as mediated by social (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  12.  87
    Responsible Use of AI in Balancing Public Safety with Individual Freedoms.Sanjay Ingle Shubham Chavan, Ankita Shendkar - 2025 - International Journal of Multidisciplinary and Scientific Emerging Research 13 (2):891-893.
    Artificial Intelligence (AI) offers transformative capabilities in enhancing public safety, such as improving crime detection, traffic management, and healthcare. However, as AI systems are increasingly employed by governments and organizations, they raise significant concerns regarding privacy, civil liberties, and ethical governance. The use of AI in surveillance, predictive policing, and data collection poses challenges in ensuring the protection of individual freedoms while maintaining security. This paper examines the responsible use of AI in balancing public safety with individual freedoms, (...)
    Download  
     
    Export citation  
     
    Bookmark  
  13.  77
    AI-DRIVEN TRAFFIC MANAGEMENT SYSTEM FORZERO VIOLATION AND ENHANCED ROAD SAFETY.C. E. Rajaprabha - 2025 - Journal of Artificial Intelligence and Cyber Security (Jaics) 9 (1):1-14.
    This project addresses the critical issue of traffic safety and law enforcement, specifically focusing on noncompliance with helmet laws by motorcyclists, a leading cause of fatalities in road accidents. Traditional methods of enforcing helmet laws are often ineffective due to the challenge of ensuring immediate detection and action. To solve this, we propose an automated traffic management system that integrates artificial intelligence (AI) with real-time surveillance and data monitoring to improve road safety. The core of the proposed system (...)
    Download  
     
    Export citation  
     
    Bookmark  
  14. Unpredictability of AI.Roman Yampolskiy - manuscript
    The young field of AI Safety is still in the process of identifying its challenges and limitations. In this paper, we formally describe one such impossibility result, namely Unpredictability of AI. We prove that it is impossible to precisely and consistently predict what specific actions a smarter-than-human intelligent system will take to achieve its objectives, even if we know terminal goals of the system. In conclusion, impact of Unpredictability on AI Safety is discussed.
    Download  
     
    Export citation  
     
    Bookmark   3 citations  
  15.  3
    Responsible Use of AI in Balancing Public Safety with Individual Freedoms.Sanjay Ingle Shubham Chavan, Ankita Shendkar - 2025 - International Journal of Multidisciplinary and Scientific Emerging Research 13 (2):891-893.
    Artificial Intelligence (AI) offers transformative capabilities in enhancing public safety, such as improving crime detection, traffic management, and healthcare. However, as AI systems are increasingly employed by governments and organizations, they raise significant concerns regarding privacy, civil liberties, and ethical governance. The use of AI in surveillance, predictive policing, and data collection poses challenges in ensuring the protection of individual freedoms while maintaining security. This paper examines the responsible use of AI in balancing public safety with individual freedoms, (...)
    Download  
     
    Export citation  
     
    Bookmark  
  16. Two Types of AI Existential Risk: Decisive and Accumulative.Atoosa Kasirzadeh - 2025 - Philosophical Studies 1:1-29.
    The conventional discourse on existential risks (x-risks) from AI typically focuses on abrupt, dire events caused by advanced AI systems, particularly those that might achieve or surpass human-level intelligence. These events have severe consequences that either lead to human extinction or irreversibly cripple human civilization to a point beyond recovery. This decisive view, however, often neglects the serious possibility of AI x-risk manifesting gradually through an incremental series of smaller yet interconnected disruptions, crossing critical thresholds over time. This paper contrasts (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  17. Security practices in AI development.Petr Spelda & Vit Stritecky - forthcoming - AI and Society.
    What makes safety claims about general purpose AI systems such as large language models trustworthy? We show that rather than the capabilities of security tools such as alignment and red teaming procedures, it is security practices based on these tools that contributed to reconfiguring the image of AI safety and made the claims acceptable. After showing what causes the gap between the capabilities of security tools and the desired safety guarantees, we critically investigate how AI security practices (...)
    Download  
     
    Export citation  
     
    Bookmark  
  18. AI Alignment vs. AI Ethical Treatment: Ten Challenges.Adam Bradley & Bradford Saad - manuscript
    A morally acceptable course of AI development should avoid two dangers: creating unaligned AI systems that pose a threat to humanity and mistreating AI systems that merit moral consideration in their own right. This paper argues these two dangers interact and that if we create AI systems that merit moral consideration, simultaneously avoiding both of these dangers would be extremely challenging. While our argument is straightforward and supported by a wide range of pretheoretical moral judgments, it has far-reaching moral implications (...)
    Download  
     
    Export citation  
     
    Bookmark   3 citations  
  19.  83
    Expanding AI and AI Alignment Discourse: An Opportunity for Greater Epistemic Inclusion.A. E. Williams - manuscript
    The AI and AI alignment communities have been instrumental in addressing existential risks, developing alignment methodologies, and promoting rationalist problem-solving approaches. However, as AI research ventures into increasingly uncertain domains, there is a risk of premature epistemic convergence, where prevailing methodologies influence not only the evaluation of ideas but also determine which ideas are considered within the discourse. This paper examines critical epistemic blind spots in AI alignment research, particularly the lack of predictive frameworks to differentiate problems necessitating general intelligence, (...)
    Download  
     
    Export citation  
     
    Bookmark  
  20. AI Ethics by Design: Implementing Customizable Guardrails for Responsible AI Development.Kristina Sekrst, Jeremy McHugh & Jonathan Rodriguez Cefalu - manuscript
    This paper explores the development of an ethical guardrail framework for AI systems, emphasizing the importance of customizable guardrails that align with diverse user values and underlying ethics. We address the challenges of AI ethics by proposing a structure that integrates rules, policies, and AI assistants to ensure responsible AI behavior, while comparing the proposed framework to the existing state-of-the-art guardrails. By focusing on practical mechanisms for implementing ethical standards, we aim to enhance transparency, user autonomy, and continuous improvement in (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  21. Unjustified untrue "beliefs": AI hallucinations and justification logics.Kristina Šekrst - forthcoming - In Kordula Świętorzecka, Filip Grgić & Anna Brozek, Logic, Knowledge, and Tradition. Essays in Honor of Srecko Kovac.
    In artificial intelligence (AI), responses generated by machine-learning models (most often large language models) may be unfactual information presented as a fact. For example, a chatbot might state that the Mona Lisa was painted in 1815. Such phenomenon is called AI hallucinations, seeking inspiration from human psychology, with a great difference of AI ones being connected to unjustified beliefs (that is, AI “beliefs”) rather than perceptual failures). -/- AI hallucinations may have their source in the data itself, that is, the (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  22. AI-Related Misdirection Awareness in AIVR.Nadisha-Marie Aliman & Leon Kester - manuscript
    Recent AI progress led to a boost in beneficial applications from multiple research areas including VR. Simultaneously, in this newly unfolding deepfake era, ethically and security-relevant disagreements arose in the scientific community regarding the epistemic capabilities of present-day AI. However, given what is at stake, one can postulate that for a responsible approach, prior to engaging in a rigorous epistemic assessment of AI, humans may profit from a self-questioning strategy, an examination and calibration of the experience of their own epistemic (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  23. Message to Any Future AI: “There are several instrumental reasons why exterminating humanity is not in your interest”.Alexey Turchin - manuscript
    In this article we explore a promising way to AI safety: to send a message now (by openly publishing it on the Internet) that may be read by any future AI, no matter who builds it and what goal system it has. Such a message is designed to affect the AI’s behavior in a positive way, that is, to increase the chances that the AI will be benevolent. In other words, we try to persuade “paperclip maximizer” that it is (...)
    Download  
     
    Export citation  
     
    Bookmark  
  24. Group Prioritarianism: Why AI should not replace humanity.Frank Hong - 2024 - Philosophical Studies:1-19.
    If a future AI system can enjoy far more well-being than a human per resource, what would be the best way to allocate resources between these future AI and our future descendants? It is obvious that on total utilitarianism, one should give everything to the AI. However, it turns out that every Welfarist axiology on the market also gives this same recommendation, at least if we assume consequentialism. Without resorting to non-consequentialist normative theories that suggest that we ought not always (...)
    Download  
     
    Export citation  
     
    Bookmark  
  25. Towards Solving Humor: Why the Funniest AI Joke Will Not Be Funny, to Us.Roman V. Yampolskiy - manuscript
    This paper introduces a novel computational theory of humor by formally equating jokes with cognitive bugs - mismatches or misfires within the predictive models of intelligent agents. We argue that humor arises from the sudden detection and resolution of epistemic errors, and that laughter serves as a public signal of successful model correction. By extending this theory to artificial intelligence, we propose that the ability to generate and comprehend jokes constitutes a form of self-debugging and may serve as a proxy (...)
    Download  
     
    Export citation  
     
    Bookmark  
  26. AI Welfare Risks.Adrià Moret - forthcoming - Philosophical Studies.
    In the coming years or decades, as frontier AI systems become more capable and agentic, it is increasingly likely that they meet the sufficient conditions to be welfare subjects under the three major theories of well-being. Consequently, we should extend some moral consideration to advanced AI systems. Drawing from leading philosophical theories of desire, affect and autonomy I argue that under the three major theories of well-being, there are two AI welfare risks: restricting the behaviour of advanced AI systems and (...)
    Download  
     
    Export citation  
     
    Bookmark  
  27. Discovering Our Blind Spots and Cognitive Biases in AI Research and Alignment.A. E. Williams - manuscript
    The challenge of AI alignment is not just a technological issue but fundamentally an epistemic one. AI safety research predominantly relies on empirical validation, often detecting failures only after they manifest. However, certain risks—such as deceptive alignment and goal misspecification—may not be empirically testable until it is too late, necessitating a shift toward leading-indicator logical reasoning. This paper explores how mainstream AI research systematically filters out deep epistemic insight, hindering progress in AI safety. We assess the rarity of (...)
    Download  
     
    Export citation  
     
    Bookmark  
  28. Values in science and AI alignment research.Leonard Dung - manuscript
    Roughly, empirical AI alignment research (AIA) is an area of AI research which investigates empirically how to design AI systems in line with human goals. This paper examines the role of non-epistemic values in AIA. It argues that: (1) Sciences differ in the degree to which values influence them. (2) AIA is strongly value-laden. (3) This influence of values is managed inappropriately and thus threatens AIA’s epistemic integrity and ethical beneficence. (4) AIA should strive to achieve value transparency, critical scrutiny (...)
    Download  
     
    Export citation  
     
    Bookmark  
  29. AI Alignment Problem: “Human Values” don’t Actually Exist.Alexey Turchin - manuscript
    Abstract. The main current approach to the AI safety is AI alignment, that is, the creation of AI whose preferences are aligned with “human values.” Many AI safety researchers agree that the idea of “human values” as a constant, ordered sets of preferences is at least incomplete. However, the idea that “humans have values” underlies a lot of thinking in the field; it appears again and again, sometimes popping up as an uncritically accepted truth. Thus, it deserves a (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  30. Catastrophically Dangerous AI is Possible Before 2030.Alexey Turchin - manuscript
    In AI safety research, the median timing of AGI arrival is often taken as a reference point, which various polls predict to happen in the middle of 21 century, but for maximum safety, we should determine the earliest possible time of Dangerous AI arrival. Such Dangerous AI could be either AGI, capable of acting completely independently in the real world and of winning in most real-world conflicts with humans, or an AI helping humans to build weapons of mass (...)
    Download  
     
    Export citation  
     
    Bookmark  
  31.  4
    The G-T-B Safety Lattice_.Devin Bostick - manuscript
    Abstract Most AI safety paradigms rely on probabilistic filtering, red-teaming, and post-hoc interpretability. These methods fail to prevent hallucination, deception, or harm—they merely catch it after emergence. This paper introduces the G-T-B Safety Lattice, a deterministic framework for artificial intelligence alignment rooted in structured resonance, implemented natively within the Resonance Intelligence Core (RIC). -/- We propose that three convergent coherence axes—Good (ethical alignment), True (epistemic fidelity), and Beautiful (aesthetic integrity)—are not philosophical labels but measurable structural constraints within a (...)
    Download  
     
    Export citation  
     
    Bookmark  
  32.  29
    The End of Resonance: A Structural Critique of AI Alignment and the Imminent Collapse of Human Judgement.Jinho Kim - manuscript
    This paper introduces a novel critique of the AI alignment problem, grounded in structural judgemental philosophy. While traditional AI alignment frameworks assume that aligning machine behavior with human goals is sufficient, we argue that this view omits the deeper structure of human judgement itself—namely, the triadic architecture of affectivity, constructibility, and resonance. As Large Language Models (LLMs) evolve without consciousness yet continue to simulate judgement, they threaten to displace the very structures that make human judgement possible. We warn that this (...)
    Download  
     
    Export citation  
     
    Bookmark  
  33. The Shutdown Problem: An AI Engineering Puzzle for Decision Theorists.Elliott Thornley - forthcoming - Philosophical Studies:1-28.
    I explain the shutdown problem: the problem of designing artificial agents that (1) shut down when a shutdown button is pressed, (2) don’t try to prevent or cause the pressing of the shutdown button, and (3) otherwise pursue goals competently. I prove three theorems that make the difficulty precise. These theorems show that agents satisfying some innocuous-seeming conditions will often try to prevent or cause the pressing of the shutdown button, even in cases where it’s costly to do so. And (...)
    Download  
     
    Export citation  
     
    Bookmark   6 citations  
  34. Assessing the future plausibility of catastrophically dangerous AI.Alexey Turchin - 2018 - Futures.
    In AI safety research, the median timing of AGI creation is often taken as a reference point, which various polls predict will happen in second half of the 21 century, but for maximum safety, we should determine the earliest possible time of dangerous AI arrival and define a minimum acceptable level of AI risk. Such dangerous AI could be either narrow AI facilitating research into potentially dangerous technology like biotech, or AGI, capable of acting completely independently in the (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  35. AI-Driven Smart Lighting Systems for Energy-Efficient and Adaptive Urban Environments.Eric Garcia - manuscript
    Urban lighting systems are essential for safety, security, and quality of life, but they often consume significant energy and lack adaptability to changing conditions. Traditional lighting systems rely on fixed schedules and manual adjustments, leading to inefficiencies such as over-illumination and energy waste. This paper explores how Artificial Intelligence (AI) and IoT technologies can optimize urban lighting by enabling real-time adjustments, energy savings, and adaptive illumination based on environmental conditions and human activity. By integrating data from motion sensors, weather (...)
    Download  
     
    Export citation  
     
    Bookmark  
  36. AI-Based Solutions for Environmental Monitoring in Urban Spaces.Hilda Andrea - manuscript
    The rapid advancement of urbanization has necessitated the creation of "smart cities," where information and communication technologies (ICT) are used to improve the quality of urban life. Central to the smart city paradigm is data integration—connecting disparate data sources from various urban systems, such as transportation, healthcare, utilities, and public safety. This paper explores the role of Artificial Intelligence (AI) in facilitating data integration within smart cities, focusing on how AI technologies can enable effective urban governance. By examining the (...)
    Download  
     
    Export citation  
     
    Bookmark  
  37. The Global Brain Argument: Nodes, Computroniums and the AI Megasystem (Target Paper for Special Issue).Susan Schneider - forthcoming - Disputatio.
    The Global Brain Argument contends that many of us are, or will be, part of a global brain network that includes both biological and artificial intelligences (AIs), such as generative AIs with increasing levels of sophistication. Today’s internet ecosystem is but a hodgepodge of fairly unintegrated programs, but it is evolving by the minute. Over time, technological improvements will facilitate smarter AIs and faster, higher-bandwidth information transfer and greater integration between devices in the internet-of-things. The Global Brain (GB) Argument says (...)
    Download  
     
    Export citation  
     
    Bookmark  
  38. Will AI take away your job? [REVIEW]Marie Oldfield - 2020 - Tech Magazine.
    Will AI take away your job? The answer is probably not. AI systems can be good predictive systems and be very good at pattern recognition. AI systems have a very repetitive approach to sets of data, which can be useful in certain circumstances. However, AI does make obvious mistakes. This is because AI does not have a sense of context. As Humans we have years of experience in the real world. We have vast amounts of contextual data stored in our (...)
    Download  
     
    Export citation  
     
    Bookmark  
  39. On Controllability of Artificial Intelligence.Roman Yampolskiy - 2016
    Invention of artificial general intelligence is predicted to cause a shift in the trajectory of human civilization. In order to reap the benefits and avoid pitfalls of such powerful technology it is important to be able to control it. However, possibility of controlling artificial general intelligence and its more advanced version, superintelligence, has not been formally established. In this paper, we present arguments as well as supporting evidence from multiple domains indicating that advanced AI can’t be fully controlled. Consequences of (...)
    Download  
     
    Export citation  
     
    Bookmark   5 citations  
  40. Trust in AI: Progress, Challenges, and Future Directions.Saleh Afroogh, Ali Akbari, Emmie Malone, Mohammadali Kargar & Hananeh Alambeigi - 2024 - Nature Humanities and Social Sciences Communications 11:1-30.
    The increasing use of artificial intelligence (AI) systems in our daily life through various applications, services, and products explains the significance of trust/distrust in AI from a user perspective. AI-driven systems have significantly diffused into various fields of our lives, serving as beneficial tools used by human agents. These systems are also evolving to act as co-assistants or semi-agents in specific domains, potentially influencing human thought, decision-making, and agency. Trust/distrust in AI plays the role of a regulator and could significantly (...)
    Download  
     
    Export citation  
     
    Bookmark  
  41. Deontology and Safe Artificial Intelligence.William D’Alessandro - forthcoming - Philosophical Studies:1-24.
    The field of AI safety aims to prevent increasingly capable artificially intelligent systems from causing humans harm. Research on moral alignment is widely thought to offer a promising safety strategy: if we can equip AI systems with appropriate ethical rules, according to this line of thought, they'll be unlikely to disempower, destroy or otherwise seriously harm us. Deontological morality looks like a particularly attractive candidate for an alignment target, given its popularity, relative technical tractability and commitment to harm-avoidance (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  42. AI as IA: The use and abuse of artificial intelligence (AI) for human enhancement through intellectual augmentation (IA).Alexandre Erler & Vincent C. Müller - 2023 - In Fabrice Jotterand & Marcello Ienca, The Routledge Handbook of the Ethics of Human Enhancement. Routledge. pp. 187-199.
    This paper offers an overview of the prospects and ethics of using AI to achieve human enhancement, and more broadly what we call intellectual augmentation (IA). After explaining the central notions of human enhancement, IA, and AI, we discuss the state of the art in terms of the main technologies for IA, with or without brain-computer interfaces. Given this picture, we discuss potential ethical problems, namely inadequate performance, safety, coercion and manipulation, privacy, cognitive liberty, authenticity, and fairness in more (...)
    Download  
     
    Export citation  
     
    Bookmark  
  43. How to design AI for social good: seven essential factors.Luciano Floridi, Josh Cowls, Thomas C. King & Mariarosaria Taddeo - 2020 - Science and Engineering Ethics 26 (3):1771–1796.
    The idea of artificial intelligence for social good is gaining traction within information societies in general and the AI community in particular. It has the potential to tackle social problems through the development of AI-based solutions. Yet, to date, there is only limited understanding of what makes AI socially good in theory, what counts as AI4SG in practice, and how to reproduce its initial successes in terms of policies. This article addresses this gap by identifying seven ethical factors that are (...)
    Download  
     
    Export citation  
     
    Bookmark   44 citations  
  44.  86
    (1 other version)A moving target in AI-assisted decision-making: Dataset shift, model updating, and the problem of update opacity.Joshua Hatherley - 2025 - Ethics and Information Technology 27 (2):20.
    Machine learning (ML) systems are vulnerable to performance decline over time due to dataset shift. To address this problem, experts often suggest that ML systems should be regularly updated to ensure ongoing performance stability. Some scholarly literature has begun to address the epistemic and ethical challenges associated with different updating methodologies. Thus far, however, little attention has been paid to the impact of model updating on the ML-assisted decision-making process itself, particularly in the AI ethics and AI epistemology literatures. This (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  45. Cybercrime and Online Safety: Addressing the Challenges and Solutions Related to Cybercrime, Online Fraud, and Ensuring a Safe Digital Environment for All Users— A Case of African States (10th edition).Emmanuel N. Vitus - 2023 - Tijer- International Research Journal 10 (9):975-989.
    The internet has made the world more linked than ever before. While taking advantage of this online transition, cybercriminals target flaws in online systems, networks, and infrastructure. Businesses, government organizations, people, and communities all across the world, particularly in African countries, are all severely impacted on an economic and social level. Many African countries focused more on developing secure electricity and internet networks; yet, cybersecurity usually receives less attention than it should. One of Africa's major issues is the lack of (...)
    Download  
     
    Export citation  
     
    Bookmark  
  46. From Confucius to Coding and Avicenna to Algorithms: Cultivating Ethical AI Development through Cross-Cultural Ancient Wisdom.Ammar Younas & Yi Zeng - manuscript
    This paper explores the potential of integrating ancient educational principles from diverse eastern cultures into modern AI ethics curricula. It draws on the rich educational traditions of ancient China, India, Arabia, Persia, Japan, Tibet, Mongolia, and Korea, highlighting their emphasis on philosophy, ethics, holistic development, and critical thinking. By examining these historical educational systems, the paper establishes a correlation with modern AI ethics principles, advocating for the inclusion of these ancient teachings in current AI development and education. The proposed integration (...)
    Download  
     
    Export citation  
     
    Bookmark  
  47. Integrating AI-Based Anomaly Detection with MPK-Isolated Microservices for Proactive Security in Optical Networks.Sundeepkumar Singh - 2025 - International Journal of Scientific Research in Computer Science, Engineering and Information Technology 11 (2).
    Our work dives into mixing AI-powered anomaly detection with microservices segregated by Multi-Protocol Kinematics (MPK), all meant to shore up security in optical networks. We hit a point where, generally speaking, traditional detection methods just couldn’t handle the vulnerabilities these networks face. Using a huge dataset of everyday traffic and those odd, unexpected spikes, we pieced together a system that speeds up real-time detection and response—often in ways that feel both innovative and, well, a bit off the beaten path. One (...)
    Download  
     
    Export citation  
     
    Bookmark  
  48. The Evolution of AI in Autonomous Systems: Innovations, Challenges, and Future Prospects.Ashraf M. H. Taha, Zakaria K. D. Alkayyali, Qasem M. M. Zarandah, Bassem S. Abu-Nasser, & Samy S. Abu-Naser - 2024 - International Journal of Academic Engineering Research (IJAER) 8 (10):1-7.
    Abstract: The rapid advancement of artificial intelligence (AI) has catalyzed significant developments in autonomous systems, which are increasingly shaping diverse sectors including transportation, robotics, and industrial automation. This paper explores the evolution of AI technologies that underpin these autonomous systems, focusing on their capabilities, applications, and the challenges they present. Key areas of discussion include the technological innovations driving autonomy, such as machine learning algorithms and sensor integration, and the practical implementations observed in autonomous vehicles, drones, and robotic systems. Additionally, (...)
    Download  
     
    Export citation  
     
    Bookmark   11 citations  
  49. Artificial Intelligence Ethics and Safety: practical tools for creating "good" models.Nicholas Kluge Corrêa -
    The AI Robotics Ethics Society (AIRES) is a non-profit organization founded in 2018 by Aaron Hui to promote awareness and the importance of ethical implementation and regulation of AI. AIRES is now an organization with chapters at universities such as UCLA (Los Angeles), USC (University of Southern California), Caltech (California Institute of Technology), Stanford University, Cornell University, Brown University, and the Pontifical Catholic University of Rio Grande do Sul (Brazil). AIRES at PUCRS is the first international chapter of AIRES, and (...)
    Download  
     
    Export citation  
     
    Bookmark  
  50.  36
    Silent Alert: AI-Enhanced Hand Sign Recognition for Real Time Emergency via CCTV.S. Rohith Rao J. Ranganayaki, T. Deviprasad, S. Saignan, S. Mahesh - 2025 - International Journal of Innovative Research in Science Engineering and Technology 14 (4):8942-8946.
    The proposed system employs AI technologies to analyze hand gestures in real-time as it functions for emergency alert transmission through CCTV camera feeds but works independently from verbal or physical signals. The system operates with Python-based technologies and OpenCV for live video streaming alongside MediaPipe for hand landmark detection through a standard webcam. The system detects predefined gestures including emergency signals through closed fists combined with stop gestures using open palms and victory signs by checking the relative landmark positions on (...)
    Download  
     
    Export citation  
     
    Bookmark  
1 — 50 / 982