Switch to: References

Add citations

You must login to add citations.
  1. On the creativity of large language models.Giorgio Franceschelli & Mirco Musolesi - forthcoming - AI and Society:1-11.
    Large language models (LLMs) are revolutionizing several areas of Artificial Intelligence. One of the most remarkable applications is creative writing, e.g., poetry or storytelling: the generated outputs are often of astonishing quality. However, a natural question arises: can LLMs be really considered creative? In this article, we first analyze the development of LLMs under the lens of creativity theories, investigating the key open questions and challenges. In particular, we focus our discussion on the dimensions of value, novelty, and surprise as (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • Chatting with Bots: AI, Speech-Acts, and the Edge of Assertion.Iwan Williams & Tim Bayne - forthcoming - Inquiry: An Interdisciplinary Journal of Philosophy.
    This paper addresses the question of whether large language model-powered chatbots are capable of assertion. According to what we call the Thesis of Chatbot Assertion (TCA), chatbots are the kinds of things that can assert, and at least some of the output produced by current-generation chatbots qualifies as assertion. We provide some motivation for TCA, arguing that it ought to be taken seriously and not simply dismissed. We also review recent objections to TCA, arguing that these objections are weighty. We (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • Human presencing: an alternative perspective on human embodiment and its implications for technology.Marie-Theres Fester-Seeger - forthcoming - AI and Society:1-19.
    Human presencing explores how people’s past encounters with others shape their present actions. In this paper, I present an alternative perspective on human embodiment in which the re-evoking of the absent can be traced to the intricate interplay of bodily dynamics. By situating the phenomenon within distributed, embodied, and dialogic approaches to language and cognition, I am overcoming the theoretical and methodological challenges involved in perceiving and acting upon what is not perceptually present. In a case study, I present strong (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • Event Knowledge in Large Language Models: The Gap Between the Impossible and the Unlikely.Carina Kauf, Anna A. Ivanova, Giulia Rambelli, Emmanuele Chersoni, Jingyuan Selena She, Zawad Chowdhury, Evelina Fedorenko & Alessandro Lenci - 2023 - Cognitive Science 47 (11):e13386.
    Word co‐occurrence patterns in language corpora contain a surprising amount of conceptual knowledge. Large language models (LLMs), trained to predict words in context, leverage these patterns to achieve impressive performance on diverse semantic tasks requiring world knowledge. An important but understudied question about LLMs’ semantic abilities is whether they acquire generalized knowledge of common events. Here, we test whether five pretrained LLMs (from 2018's BERT to 2023's MPT) assign a higher likelihood to plausible descriptions of agent−patient interactions than to minimally (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • Assessing the Strengths and Weaknesses of Large Language Models.Shalom Lappin - 2023 - Journal of Logic, Language and Information 33 (1):9-20.
    The transformers that drive chatbots and other AI systems constitute large language models (LLMs). These are currently the focus of a lively discussion in both the scientific literature and the popular media. This discussion ranges from hyperbolic claims that attribute general intelligence and sentience to LLMs, to the skeptical view that these devices are no more than “stochastic parrots”. I present an overview of some of the weak arguments that have been presented against LLMs, and I consider several of the (...)
    Download  
     
    Export citation  
     
    Bookmark   7 citations  
  • Does thought require sensory grounding? From pure thinkers to large language models.David J. Chalmers - 2023 - Proceedings and Addresses of the American Philosophical Association 97:22-45.
    Does the capacity to think require the capacity to sense? A lively debate on this topic runs throughout the history of philosophy and now animates discussions of artificial intelligence. Many have argued that AI systems such as large language models cannot think and understand if they lack sensory grounding. I argue that thought does not require sensory grounding: there can be pure thinkers who can think without any sensory capacities. As a result, the absence of sensory grounding does not entail (...)
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  • Real Sparks of Artificial Intelligence and the Importance of Inner Interpretability.Alex Grzankowski - forthcoming - Inquiry: An Interdisciplinary Journal of Philosophy.
    The present paper looks at one of the most thorough articles on the intelligence of GPT, research conducted by engineers at Microsoft. Although there is a great deal of value in their work, I will argue that, for familiar philosophical reasons, their methodology, ‘Black-box Interpretability’ is wrongheaded. But there is a better way. There is an exciting and emerging discipline of ‘Inner Interpretability’ (also sometimes called ‘White-box Interpretability’) that aims to uncover the internal activations and weights of models in order (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • Trust, understanding, and machine translation: the task of translation and the responsibility of the translator.Melvin Chen - 2024 - AI and Society 39 (5):2307-2319.
    Could translation be fully automated? We must first acknowledge the complexity, ambiguity, and diversity of natural languages. These aspects of natural languages, when combined with a particular dilemma known as the computational dilemma, appear to imply that the machine translator faces certain obstacles that a human translator has already managed to overcome. At the same time, science has not yet solved the problem of how human brains process natural languages and how human beings come to acquire natural language understanding. We (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • A Loosely Wittgensteinian Conception of the Linguistic Understanding of Large Language Models like BERT, GPT-3, and ChatGPT.Reto Gubelmann - 2023 - Grazer Philosophische Studien 99 (4):485-523.
    In this article, I develop a loosely Wittgensteinian conception of what it takes for a being, including an AI system, to understand language, and I suggest that current state of the art systems are closer to fulfilling these requirements than one might think. Developing and defending this claim has both empirical and conceptual aspects. The conceptual aspects concern the criteria that are reasonably applied when judging whether some being understands language; the empirical aspects concern the question whether a given being (...)
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  • Still no lie detector for language models: probing empirical and conceptual roadblocks.Benjamin A. Levinstein & Daniel A. Herrmann - forthcoming - Philosophical Studies:1-27.
    We consider the questions of whether or not large language models (LLMs) have beliefs, and, if they do, how we might measure them. First, we consider whether or not we should expect LLMs to have something like beliefs in the first place. We consider some recent arguments aiming to show that LLMs cannot have beliefs. We show that these arguments are misguided. We provide a more productive framing of questions surrounding the status of beliefs in LLMs, and highlight the empirical (...)
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  • The argument for near-term human disempowerment through AI.Leonard Dung - 2024 - AI and Society:1-14.
    Many researchers and intellectuals warn about extreme risks from artificial intelligence. However, these warnings typically came without systematic arguments in support. This paper provides an argument that AI will lead to the permanent disempowerment of humanity, e.g. human extinction, by 2100. It rests on four substantive premises which it motivates and defends: first, the speed of advances in AI capability, as well as the capability level current systems have already reached, suggest that it is practically possible to build AI systems (...)
    Download  
     
    Export citation  
     
    Bookmark   4 citations  
  • Grounding the Vector Space of an Octopus: Word Meaning from Raw Text.Anders Søgaard - 2021 - Minds and Machines 33 (1):33-54.
    Most, if not all, philosophers agree that computers cannot learn what words refers to from raw text alone. While many attacked Searle’s Chinese Room thought experiment, no one seemed to question this most basic assumption. For how can computers learn something that is not in the data? Emily Bender and Alexander Koller ( 2020 ) recently presented a related thought experiment—the so-called Octopus thought experiment, which replaces the rule-based interlocutor of Searle’s thought experiment with a neural language model. The Octopus (...)
    Download  
     
    Export citation  
     
    Bookmark   3 citations  
  • Understanding models understanding language.Anders Søgaard - 2022 - Synthese 200 (6):1-16.
    Landgrebe and Smith :2061–2081, 2021) present an unflattering diagnosis of recent advances in what they call language-centric artificial intelligence—perhaps more widely known as natural language processing: The models that are currently employed do not have sufficient expressivity, will not generalize, and are fundamentally unable to induce linguistic semantics, they say. The diagnosis is mainly derived from an analysis of the widely used Transformer architecture. Here I address a number of misunderstandings in their analysis, and present what I take to be (...)
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  • Personhood and AI: Why large language models don’t understand us.Jacob Browning - 2023 - AI and Society 39 (5):2499-2506.
    Recent artificial intelligence advances, especially those of large language models (LLMs), have increasingly shown glimpses of human-like intelligence. This has led to bold claims that these systems are no longer a mere “it” but now a “who,” a kind of person deserving respect. In this paper, I argue that this view depends on a Cartesian account of personhood, on which identifying someone as a person is based on their cognitive sophistication and ability to address common-sense reasoning problems. I contrast this (...)
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  • Automated inauthenticity.Mark Ressler - forthcoming - AI and Society.
    Large language models and other generative artificial intelligence systems are achieving increasingly impressive results, though the quality of those results still seems dull and uninspired. This paper argues that this poor quality can be linked to the philosophical notion of inauthenticity as presented by Kierkegaard, Nietzsche, and Heidegger, and that this inauthenticity is fundamentally grounded in the design and structure of such systems by virtue of the way they statistically level down the materials on which they are trained. Although it (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • Negotiating becoming: a Nietzschean critique of large language models.Simon W. S. Fischer & Bas de Boer - 2024 - Ethics and Information Technology 26 (3):1-12.
    Large language models (LLMs) structure the linguistic landscape by reflecting certain beliefs and assumptions. In this paper, we address the risk of people unthinkingly adopting and being determined by the values or worldviews embedded in LLMs. We provide a Nietzschean critique of LLMs and, based on the concept of will to power, consider LLMs as will-to-power organisations. This allows us to conceptualise the interaction between self and LLMs as power struggles, which we understand as negotiation. Currently, the invisibility and incomprehensibility (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • Bayesian Surprise Predicts Human Event Segmentation in Story Listening.Manoj Kumar, Ariel Goldstein, Sebastian Michelmann, Jeffrey M. Zacks, Uri Hasson & Kenneth A. Norman - 2023 - Cognitive Science 47 (10):e13343.
    Event segmentation theory posits that people segment continuous experience into discrete events and that event boundaries occur when there are large transient increases in prediction error. Here, we set out to test this theory in the context of story listening, by using a deep learning language model (GPT‐2) to compute the predicted probability distribution of the next word, at each point in the story. For three stories, we used the probability distributions generated by GPT‐2 to compute the time series of (...)
    Download  
     
    Export citation  
     
    Bookmark  
  • LMMS reloaded: Transformer-based sense embeddings for disambiguation and beyond.Daniel Loureiro, Alípio Mário Jorge & Jose Camacho-Collados - 2022 - Artificial Intelligence 305 (C):103661.
    Download  
     
    Export citation  
     
    Bookmark  
  • Machine learning in human creativity: status and perspectives.Mirko Farina, Andrea Lavazza, Giuseppe Sartori & Witold Pedrycz - forthcoming - AI and Society:1-13.
    As we write this research paper, we notice an explosion in popularity of machine learning in numerous fields (ranging from governance, education, and management to criminal justice, fraud detection, and internet of things). In this contribution, rather than focusing on any of those fields, which have been well-reviewed already, we decided to concentrate on a series of more recent applications of deep learning models and technologies that have only recently gained significant track in the relevant literature. These applications are concerned (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  • The Philosophising Machine – a Specification of the Turing Test.Arthur C. Schwaninger - 2022 - Philosophia 50 (3):1437-1453.
    Block’s, 5–43 1981) anti-behaviourist attack of the Turing Test not only illustrates that the test is a non-sufficient criterion for attributing thought; I suggest that it also exemplifies the limiting case of the more general concern that a machine which has access to enormous amounts of data can pass the Turing Test by simple symbol-manipulation techniques. If the answers to a human interrogator are entailed by the machines’ data, the Turing Test offers no clear criterion to distinguish between a thinking (...)
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  • Is it Possible to Preserve a Language using only Data?Joshua Bensemann, Jason Brown, Michael Witbrock & Vithya Yogarajan - 2023 - Cognitive Science 47 (6):e13300.
    Many of our spoken languages are endangered and rapidly becoming extinct. Due to this, there are attempts to preserve as many of those languages as possible. One preservation approach is combining data collection and artificial intelligence‐based language models. However, current data collection methods may only capture static data from a dynamic cognitive process. If data are not genuinely capturing the dynamic process, it raises questions about whether they capture all the essential knowledge about how a language functions. Here, we discuss (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation