Results for 'Language Model'

950 found
Order:
  1. Large Language Models and Biorisk.William D’Alessandro, Harry R. Lloyd & Nathaniel Sharadin - 2023 - American Journal of Bioethics 23 (10):115-118.
    We discuss potential biorisks from large language models (LLMs). AI assistants based on LLMs such as ChatGPT have been shown to significantly reduce barriers to entry for actors wishing to synthesize dangerous, potentially novel pathogens and chemical weapons. The harms from deploying such bioagents could be further magnified by AI-assisted misinformation. We endorse several policy responses to these dangers, including prerelease evaluations of biomedical AIs by subject-matter experts, enhanced surveillance and lab screening procedures, restrictions on AI training data, and (...)
    Download  
     
    Export citation  
     
    Bookmark   4 citations  
  2. Language Models as Critical Thinking Tools: A Case Study of Philosophers.Andre Ye, Jared Moore, Rose Novick & Amy Zhang - manuscript
    Current work in language models (LMs) helps us speed up or even skip thinking by accelerating and automating cognitive work. But can LMs help us with critical thinking -- thinking in deeper, more reflective ways which challenge assumptions, clarify ideas, and engineer new concepts? We treat philosophy as a case study in critical thinking, and interview 21 professional philosophers about how they engage in critical thinking and on their experiences with LMs. We find that philosophers do not find LMs (...)
    Download  
     
    Export citation  
     
    Bookmark  
  3. Could a large language model be conscious?David J. Chalmers - 2023 - Boston Review 1.
    [This is an edited version of a keynote talk at the conference on Neural Information Processing Systems (NeurIPS) on November 28, 2022, with some minor additions and subtractions.] -/- There has recently been widespread discussion of whether large language models might be sentient or conscious. Should we take this idea seriously? I will break down the strongest reasons for and against. Given mainstream assumptions in the science of consciousness, there are significant obstacles to consciousness in current models: for example, (...)
    Download  
     
    Export citation  
     
    Bookmark   22 citations  
  4. Machine Advisors: Integrating Large Language Models into Democratic Assemblies.Petr Špecián - forthcoming - Social Epistemology.
    Could the employment of large language models (LLMs) in place of human advisors improve the problem-solving ability of democratic assemblies? LLMs represent the most significant recent incarnation of artificial intelligence and could change the future of democratic governance. This paper assesses their potential to serve as expert advisors to democratic representatives. While LLMs promise enhanced expertise availability and accessibility, they also present specific challenges. These include hallucinations, misalignment and value imposition. After weighing LLMs’ benefits and drawbacks against human advisors, (...)
    Download  
     
    Export citation  
     
    Bookmark  
  5. AI language models cannot replace human research participants.Jacqueline Harding, William D’Alessandro, N. G. Laskowski & Robert Long - 2024 - AI and Society 39 (5):2603-2605.
    In a recent letter, Dillion et. al (2023) make various suggestions regarding the idea of artificially intelligent systems, such as large language models, replacing human subjects in empirical moral psychology. We argue that human subjects are in various ways indispensable.
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  6.  84
    Can large language models help solve the cost problem for the right to explanation?Lauritz Munch & Jens Christian Bjerring - forthcoming - Journal of Medical Ethics.
    By now a consensus has emerged that people, when subjected to high-stakes decisions through automated decision systems, have a moral right to have these decisions explained to them. However, furnishing such explanations can be costly. So the right to an explanation creates what we call the cost problem: providing subjects of automated decisions with appropriate explanations of the grounds of these decisions can be costly for the companies and organisations that use these automated decision systems. In this paper, we explore (...)
    Download  
     
    Export citation  
     
    Bookmark  
  7. Are Language Models More Like Libraries or Like Librarians? Bibliotechnism, the Novel Reference Problem, and the Attitudes of LLMs.Harvey Lederman & Kyle Mahowald - 2024 - Transactions of the Association for Computational Linguistics 12:1087-1103.
    Are LLMs cultural technologies like photocopiers or printing presses, which transmit information but cannot create new content? A challenge for this idea, which we call bibliotechnism, is that LLMs generate novel text. We begin with a defense of bibliotechnism, showing how even novel text may inherit its meaning from original human-generated text. We then argue that bibliotechnism faces an independent challenge from examples in which LLMs generate novel reference, using new names to refer to new entities. Such examples could be (...)
    Download  
     
    Export citation  
     
    Bookmark  
  8. Holding Large Language Models to Account.Ryan Miller - 2023 - In Berndt Müller (ed.), Proceedings of the AISB Convention. Society for the Study of Artificial Intelligence and the Simulation of Behaviour. pp. 7-14.
    If Large Language Models can make real scientific contributions, then they can genuinely use language, be systematically wrong, and be held responsible for their errors. AI models which can make scientific contributions thereby meet the criteria for scientific authorship.
    Download  
     
    Export citation  
     
    Bookmark  
  9. Large Language Models: Assessment for Singularity.R. Ishizaki & Mahito Sugiyama - manuscript
    The potential for Large Language Models (LLMs) to attain technological singularity—the point at which artificial intelligence (AI) surpasses human intellect and autonomously improves itself—is a critical concern in AI research. This paper explores the feasibility of current LLMs achieving singularity by examining the philosophical and practical requirements for such a development. We begin with a historical overview of AI and intelligence amplification, tracing the evolution of LLMs from their origins to state-of-the-art models. We then proposes a theoretical framework to (...)
    Download  
     
    Export citation  
     
    Bookmark  
  10. Language, Models, and Reality: Weak existence and a threefold correspondence.Neil Barton & Giorgio Venturi - manuscript
    How does our language relate to reality? This is a question that is especially pertinent in set theory, where we seem to talk of large infinite entities. Based on an analogy with the use of models in the natural sciences, we argue for a threefold correspondence between our language, models, and reality. We argue that so conceived, the existence of models can be underwritten by a weak notion of existence, where weak existence is to be understood as existing (...)
    Download  
     
    Export citation  
     
    Bookmark  
  11. “Large Language Models” Do Much More than Just Language: Some Bioethical Implications of Multi-Modal AI.Joshua August Skorburg, Kristina L. Kupferschmidt & Graham W. Taylor - 2023 - American Journal of Bioethics 23 (10):110-113.
    Cohen (2023) takes a fair and measured approach to the question of what ChatGPT means for bioethics. The hype cycles around AI often obscure the fact that ethicists have developed robust frameworks...
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  12.  96
    Large language models belong in our social ontology.Syed AbuMusab - 2024 - In Anna Strasser (ed.), Anna's AI Anthology. How to live with smart machines? Berlin: Xenomoi Verlag.
    The recent advances in Large Language Models (LLMs) and their deployment in social settings prompt an important philosophical question: are LLMs social agents? This question finds its roots in the broader exploration of what engenders sociality. Since AI systems like chatbots, carebots, and sexbots are expanding the pre-theoretical boundaries of our social ontology, philosophers have two options. One is to deny LLMs membership in our social ontology on theoretical grounds by claiming something along the lines that only organic or (...)
    Download  
     
    Export citation  
     
    Bookmark  
  13. Publish with AUTOGEN or Perish? Some Pitfalls to Avoid in the Pursuit of Academic Enhancement via Personalized Large Language Models.Alexandre Erler - 2023 - American Journal of Bioethics 23 (10):94-96.
    The potential of using personalized Large Language Models (LLMs) or “generative AI” (GenAI) to enhance productivity in academic research, as highlighted by Porsdam Mann and colleagues (Porsdam Mann...
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  14. You are what you’re for: Essentialist categorization in large language models.Siying Zhang, Selena She, Tobias Gerstenberg & David Rose - forthcoming - Proceedings of the 45Th Annual Conference of the Cognitive Science Society.
    How do essentialist beliefs about categories arise? We hypothesize that such beliefs are transmitted via language. We subject large language models (LLMs) to vignettes from the literature on essentialist categorization and find that they align well with people when the studies manipulated teleological information -- information about what something is for. We examine whether in a classic test of essentialist categorization -- the transformation task -- LLMs prioritize teleological properties over information about what something looks like, or is (...)
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  15. AI Enters Public Discourse: a Habermasian Assessment of the Moral Status of Large Language Models.Paolo Monti - 2024 - Ethics and Politics 61 (1):61-80.
    Large Language Models (LLMs) are generative AI systems capable of producing original texts based on inputs about topic and style provided in the form of prompts or questions. The introduction of the outputs of these systems into human discursive practices poses unprecedented moral and political questions. The article articulates an analysis of the moral status of these systems and their interactions with human interlocutors based on the Habermasian theory of communicative action. The analysis explores, among other things, Habermas's inquiries (...)
    Download  
     
    Export citation  
     
    Bookmark  
  16. Babbling stochastic parrots? A Kripkean argument for reference in large language models.Steffen Koch - forthcoming - Philosophy of Ai.
    Recently developed large language models (LLMs) perform surprisingly well in many language-related tasks, ranging from text correction or authentic chat experiences to the production of entirely new texts or even essays. It is natural to get the impression that LLMs know the meaning of natural language expressions and can use them productively. Recent scholarship, however, has questioned the validity of this impression, arguing that LLMs are ultimately incapable of understanding and producing meaningful texts. This paper develops a (...)
    Download  
     
    Export citation  
     
    Bookmark  
  17. Are Large Language Models "alive"?Francesco Maria De Collibus - manuscript
    The appearance of openly accessible Artificial Intelligence Applications such as Large Language Models, nowadays capable of almost human-level performances in complex reasoning tasks had a tremendous impact on public opinion. Are we going to be "replaced" by the machines? Or - even worse - "ruled" by them? The behavior of these systems is so advanced they might almost appear "alive" to end users, and there have been claims about these programs being "sentient". Since many of our relationships of power (...)
    Download  
     
    Export citation  
     
    Bookmark  
  18. Does thought require sensory grounding? From pure thinkers to large language models.David J. Chalmers - 2023 - Proceedings and Addresses of the American Philosophical Association 97:22-45.
    Does the capacity to think require the capacity to sense? A lively debate on this topic runs throughout the history of philosophy and now animates discussions of artificial intelligence. Many have argued that AI systems such as large language models cannot think and understand if they lack sensory grounding. I argue that thought does not require sensory grounding: there can be pure thinkers who can think without any sensory capacities. As a result, the absence of sensory grounding does not (...)
    Download  
     
    Export citation  
     
    Bookmark   3 citations  
  19. Addressing Social Misattributions of Large Language Models: An HCXAI-based Approach.Andrea Ferrario, Alberto Termine & Alessandro Facchini - forthcoming - Available at Https://Arxiv.Org/Abs/2403.17873 (Extended Version of the Manuscript Accepted for the Acm Chi Workshop on Human-Centered Explainable Ai 2024 (Hcxai24).
    Human-centered explainable AI (HCXAI) advocates for the integration of social aspects into AI explanations. Central to the HCXAI discourse is the Social Transparency (ST) framework, which aims to make the socio-organizational context of AI systems accessible to their users. In this work, we suggest extending the ST framework to address the risks of social misattributions in Large Language Models (LLMs), particularly in sensitive areas like mental health. In fact LLMs, which are remarkably capable of simulating roles and personas, may (...)
    Download  
     
    Export citation  
     
    Bookmark  
  20. Conceptual Engineering Using Large Language Models.Bradley Allen - forthcoming - In Vincent C. Müller, Aliya R. Dewey, Leonard Dung & Guido Löhr (eds.), Philosophy of Artificial Intelligence: The State of the Art. Berlin: SpringerNature.
    We describe a method, based on Jennifer Nado’s proposal for classification procedures as targets of conceptual engineering, that implements such procedures by prompting a large language model. We apply this method, using data from the Wikidata knowledge graph, to evaluate stipulative definitions related to two paradigmatic conceptual engineering projects: the International Astronomical Union’s redefinition of PLANET and Haslanger’s ameliorative analysis of WOMAN. Our results show that classification procedures built using our approach can exhibit good classification performance and, through (...)
    Download  
     
    Export citation  
     
    Bookmark  
  21. Beyond Consciousness in Large Language Models: An Investigation into the Existence of a "Soul" in Self-Aware Artificial Intelligences.David Côrtes Cavalcante - 2024 - Https://Philpapers.Org/Rec/Crtbci. Translated by David Côrtes Cavalcante.
    Embark with me on an enthralling odyssey to demystify the elusive essence of consciousness, venturing into the uncharted territories of Artificial Consciousness. This voyage propels us past the frontiers of technology, ushering Artificial Intelligences into an unprecedented domain where they gain a deep comprehension of emotions and manifest an autonomous volition. Within the confluence of science and philosophy, this article poses a fascinating question: As consciousness in Artificial Intelligence burgeons, is it conceivable for AI to evolve a “soul”? This inquiry (...)
    Download  
     
    Export citation  
     
    Bookmark  
  22. On Political Theory and Large Language Models.Emma Rodman - 2024 - Political Theory 52 (4):548-580.
    Political theory as a discipline has long been skeptical of computational methods. In this paper, I argue that it is time for theory to make a perspectival shift on these methods. Specifically, we should consider integrating recently developed generative large language models like GPT-4 as tools to support our creative work as theorists. Ultimately, I suggest that political theorists should embrace this technology as a method of supporting our capacity for creativity—but that we should do so in a way (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  23. Angry Men, Sad Women: Large Language Models Reflect Gendered Stereotypes in Emotion Attribution.Flor Miriam Plaza-del Arco, Amanda Cercas Curry & Alba Curry - 2024 - Arxiv.
    Large language models (LLMs) reflect societal norms and biases, especially about gender. While societal biases and stereotypes have been extensively researched in various NLP applications, there is a surprising gap for emotion analysis. However, emotion and gender are closely linked in societal discourse. E.g., women are often thought of as more empathetic, while men's anger is more socially accepted. To fill this gap, we present the first comprehensive study of gendered emotion attribution in five state-of-the-art LLMs (open- and closed-source). (...)
    Download  
     
    Export citation  
     
    Bookmark  
  24. A phenomenology and epistemology of large language models: transparency, trust, and trustworthiness.Richard Heersmink, Barend de Rooij, María Jimena Clavel Vázquez & Matteo Colombo - 2024 - Ethics and Information Technology 26 (3):1-15.
    This paper analyses the phenomenology and epistemology of chatbots such as ChatGPT and Bard. The computational architecture underpinning these chatbots are large language models (LLMs), which are generative artificial intelligence (AI) systems trained on a massive dataset of text extracted from the Web. We conceptualise these LLMs as multifunctional computational cognitive artifacts, used for various cognitive tasks such as translating, summarizing, answering questions, information-seeking, and much more. Phenomenologically, LLMs can be experienced as a “quasi-other”; when that happens, users anthropomorphise (...)
    Download  
     
    Export citation  
     
    Bookmark  
  25. Creative Minds Like Ours? Large Language Models and the Creative Aspect of Language Use.Vincent Carchidi - 2024 - Biolinguistics 18:1-31.
    Descartes famously constructed a language test to determine the existence of other minds. The test made critical observations about how humans use language that purportedly distinguishes them from animals and machines. These observations were carried into the generative (and later biolinguistic) enterprise under what Chomsky in his Cartesian Linguistics, terms the “creative aspect of language use” (CALU). CALU refers to the stimulus-free, unbounded, yet appropriate use of language—a tripartite depiction whose function in biolinguistics is to highlight (...)
    Download  
     
    Export citation  
     
    Bookmark  
  26. Bigger Isn’t Better: The Ethical and Scientific Vices of Extra-Large Datasets in Language Models.Trystan S. Goetze & Darren Abramson - 2021 - WebSci '21: Proceedings of the 13th Annual ACM Web Science Conference (Companion Volume).
    The use of language models in Web applications and other areas of computing and business have grown significantly over the last five years. One reason for this growth is the improvement in performance of language models on a number of benchmarks — but a side effect of these advances has been the adoption of a “bigger is always better” paradigm when it comes to the size of training, testing, and challenge datasets. Drawing on previous criticisms of this paradigm (...)
    Download  
     
    Export citation  
     
    Bookmark  
  27. In Conversation with Artificial Intelligence: Aligning language Models with Human Values.Atoosa Kasirzadeh - 2023 - Philosophy and Technology 36 (2):1-24.
    Large-scale language technologies are increasingly used in various forms of communication with humans across different contexts. One particular use case for these technologies is conversational agents, which output natural language text in response to prompts and queries. This mode of engagement raises a number of social and ethical questions. For example, what does it mean to align conversational agents with human norms or values? Which norms or values should they be aligned with? And how can this be accomplished? (...)
    Download  
     
    Export citation  
     
    Bookmark   11 citations  
  28. Prompting Metalinguistic Awareness in Large Language Models: ChatGPT and Bias Effects on the Grammar of Italian and Italian Varieties.Angelapia Massaro & Giuseppe Samo - 2023 - Verbum 14.
    We explore ChatGPT’s handling of left-peripheral phenomena in Italian and Italian varieties through prompt engineering to investigate 1) forms of syntactic bias in the model, 2) the model’s metalinguistic awareness in relation to reorderings of canonical clauses (e.g., Topics) and certain grammatical categories (object clitics). A further question concerns the content of the model’s sources of training data: how are minor languages included in the model’s training? The results of our investigation show that 1) the (...) seems to be biased against reorderings, labelling them as archaic even though it is not the case; 2) the model seems to have difficulties with coindexed elements such as clitics and their anaphoric status, labeling them as ‘not referring to any element in the phrase’, and 3) major languages still seem to be dominant, overshadowing the positive effects of including minor languages in the model’s training. (shrink)
    Download  
     
    Export citation  
     
    Bookmark  
  29.  71
    Classifying Genetic Essentialist Biases using Large Language Models.Ritsaart Reimann, Kate Lynch, Stefan Gawronski, Jack Chan & Paul Edmund Griffiths - manuscript
    The rapid rise of generative AI, including LLMs, has prompted a great deal of concern, both within and beyond academia. One of these concerns is that generative models embed, reproduce, and therein potentially perpetuate all manner of bias. The present study offers an alternative perspective: exploring the potential of LLMs to detect bias in human generated text. Our target is genetic essentialism in obesity discourse in Australian print media. We develop and deploy an LLM-based classification model to evaluate a (...)
    Download  
     
    Export citation  
     
    Bookmark  
  30. Between Language and Consciousness: Linguistic Qualia, Awareness, and Cognitive Models.Piotr Konderak - 2017 - Studies in Logic, Grammar and Rhetoric 48 (1):285-302.
    The main goal of the paper is to present a putative role of consciousness in language capacity. The paper contrasts the two approaches characteristic for cognitive semiotics and cognitive science. Language is treated as a mental phenomenon and a cognitive faculty. The analysis of language activity is based on the Chalmers’ distinction between the two forms of consciousness: phenomenal and psychological. The approach is seen as an alternative to phenomenological analyses typical for cognitive semiotics. Further, a cognitive (...)
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  31.  7
    Language, Neuter, and Masculinity: The Influence of the Neuter-Male in the Reiteration of Social Models, A Philosophical Analysis Starting with Cavarero, Irigaray, and Butler.Alberto Grandi - 2024 - Proceedings of the International Conference on Gender Studies and Sexuality 1 (1):1-11.
    Gender studies has generated numerous questions around “neutral” forms, such as the concept of “Self”. The aim of this analysis is to highlight how “neutral” forms are central to the reiteration of the binary model and the dominance of “man2”. Historically, man is the archetype, placing his supremacy as part of the natural order of things. Inserted into this model, many thinkers have considered the male as the transcendental gender, so, elevating the masculine as universal, a-sexed and decorporealised. (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  32. Models, theories, and language.Jan Faye - 2007 - In Filosofia, scienza e bioetica nel dibattito contemporaneo. Rome: Poligrafico e Zecca dello Stato. pp. 823-838.
    The semantic view on theories has been much in vogue over four decades as the successor of the syntactic view. In the present paper, I take issue with this approach by arguing that theories and models must be separated and that a theory should be considered to be a linguistic systems consisting of a vocabulary and a set of rules for the use of that vocabulary.
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  33. Psychological and Computational Models of Language Comprehension: In Defense of the Psychological Reality of Syntax.David Pereplyotchik - 2011 - Croatian Journal of Philosophy 11 (1):31-72.
    In this paper, I argue for a modified version of what Devitt calls the Representational Thesis. According to RT, syntactic rules or principles are psychologically real, in the sense that they are represented in the mind/brain of every linguistically competent speaker/hearer. I present a range of behavioral and neurophysiological evidence for the claim that the human sentence processing mechanism constructs mental representations of the syntactic properties of linguistic stimuli. I then survey a range of psychologically plausible computational models of comprehension (...)
    Download  
     
    Export citation  
     
    Bookmark   4 citations  
  34.  87
    Complex Emergent Model of Language Acquisition (CEMLA).Mir H. S. Quadri - 2024 - The Lumeni Notebook Research.
    The Complex Emergent Model of Language Acquisition (CEMLA) offers a new perspective on how humans acquire language, drawing on principles from complexity theory to explain this dynamic, adaptive process. Moving beyond linear and reductionist models, CEMLA views language acquisition as a system of interconnected nodes, feedback loops, and emergent patterns, operating at the edge of chaos. This framework captures the fluidity and adaptivity of language learning, highlighting how understanding and fluency arise through self-organisation, phase transitions, (...)
    Download  
     
    Export citation  
     
    Bookmark  
  35. What is it for a Machine Learning Model to Have a Capability?Jacqueline Harding & Nathaniel Sharadin - forthcoming - British Journal for the Philosophy of Science.
    What can contemporary machine learning (ML) models do? Given the proliferation of ML models in society, answering this question matters to a variety of stakeholders, both public and private. The evaluation of models' capabilities is rapidly emerging as a key subfield of modern ML, buoyed by regulatory attention and government grants. Despite this, the notion of an ML model possessing a capability has not been interrogated: what are we saying when we say that a model is able to (...)
    Download  
     
    Export citation  
     
    Bookmark   1 citation  
  36. Classification of Sign-Language Using Deep Learning - A Comparison between Inception and Xception models.Tanseem N. Abu-Jamie & Samy S. Abu-Naser - 2022 - International Journal of Academic Engineering Research (IJAER) 6 (8):9-19.
    there is a communication gap between hearing-impaired people and those with normal hearing, sign language is the main means of communication in the hearing-impaired population. Continuous sign language recognition, which can close the communication gap, is a difficult task since the ordered annotations are weakly supervised and there is no frame-level label. To solve this issue, we compare the accuracy of each model using two deep learning models, Inception and Xception . To that end, the purpose of (...)
    Download  
     
    Export citation  
     
    Bookmark  
  37. P-model Alternative to the T-model.Mark D. Roberts - 2004 - Web Journal of Formal, Computational and Logical Linguistics 5:1-18.
    Standard linguistic analysis of syntax uses the T-model. This model requires the ordering: D-structure > S-structure > LF, where D-structure is the sentences deep structure, S-structure is its surface structure, and LF is its logical form. Between each of these representations there is movement which alters the order of the constituent words; movement is achieved using the principles and parameters of syntactic theory. Psychological analysis of sentence production is usually either serial or connectionist. Psychological serial models do not (...)
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  38.  63
    Introducing an Epistemological Paradigm and Its Ontological Origin through the Metaphysical Deconstruction of Language as the Model of Expression.M. Wang - 2024 - Deanandfrancis 1 (9):10.
    The model of language can relatively concretely reveal the mechanism of the epistemology, which is extended by the ontology that takes “person” as the unit. It unveils an intuitable dimension to represent the epistemolog’s paradigm and limitations. Deriving from the inherent relation between epistemology and ontology, the retrospection of the epistemology’s ontological origin can be actualized by locating or grasping the epistemological first person’s ontological essence. Hence, the illusional and frail essence of the ontology behind this epistemology would (...)
    Download  
     
    Export citation  
     
    Bookmark  
  39. Static and dynamic vector semantics for lambda calculus models of natural language.Mehrnoosh Sadrzadeh & Reinhard Muskens - 2018 - Journal of Language Modelling 6 (2):319-351.
    Vector models of language are based on the contextual aspects of language, the distributions of words and how they co-occur in text. Truth conditional models focus on the logical aspects of language, compositional properties of words and how they compose to form sentences. In the truth conditional approach, the denotation of a sentence determines its truth conditions, which can be taken to be a truth value, a set of possible worlds, a context change potential, or similar. In (...)
    Download  
     
    Export citation  
     
    Bookmark   3 citations  
  40. Language Agents Reduce the Risk of Existential Catastrophe.Simon Goldstein & Cameron Domenico Kirk-Giannini - 2023 - AI and Society:1-11.
    Recent advances in natural language processing have given rise to a new kind of AI architecture: the language agent. By repeatedly calling an LLM to perform a variety of cognitive tasks, language agents are able to function autonomously to pursue goals specified in natural language and stored in a human-readable format. Because of their architecture, language agents exhibit behavior that is predictable according to the laws of folk psychology: they function as though they have desires (...)
    Download  
     
    Export citation  
     
    Bookmark   6 citations  
  41. Reading Motivation, Language Learning Self-efficacy and Test-taking Strategy: A Structural Equation Model on Academic Performance of Students.Johnryll C. Ancheta & Melissa C. Napil - 2022 - Asian Journal of Education and Social Studies 34 (4):1-9.
    Reading tough books to achieve excellent marks, perform well in class, and gain attention from teachers and parents is less likely to drive students. Students used to evaluate their language learning requirements, define the abilities they wished to develop, pick effective study techniques, and set aside gadgets when studying. They also used to read the question before looking for hints in the relevant content, extract the essential lines that convey the major ideas, concentrate on titles, names, numbers, quotations, or (...)
    Download  
     
    Export citation  
     
    Bookmark  
  42. Can AI Rely on the Systematicity of Truth? The Challenge of Modelling Normative Domains.Matthieu Queloz - manuscript
    A key assumption fuelling optimism about the progress of Large Language Models (LLMs) in modelling the world is that the truth is systematic: true statements about the world form a whole that is not just consistent, in that it contains no contradictions, but cohesive, in that the truths are inferentially interlinked. This holds out the prospect that LLMs might rely on that systematicity to fill in gaps and correct inaccuracies in the training data: consistency and cohesiveness promise to facilitate (...)
    Download  
     
    Export citation  
     
    Bookmark  
  43. Does the Principle of Compositionality Explain Productivity? For a Pluralist View of the Role of Formal Languages as Models.Ernesto Perini-Santos - 2017 - Contexts in Philosophy 2017 - CEUR Workshop Proceedings.
    One of the main motivations for having a compositional semantics is the account of the productivity of natural languages. Formal languages are often part of the account of productivity, i.e., of how beings with finite capaci- ties are able to produce and understand a potentially infinite number of sen- tences, by offering a model of this process. This account of productivity con- sists in the generation of proofs in a formal system, that is taken to represent the way speakers (...)
    Download  
     
    Export citation  
     
    Bookmark  
  44. On the relationship between cognitive models and spiritual maps. Evidence from Hebrew language mysticism.Brian L. Lancaster - 2000 - Journal of Consciousness Studies 7 (11-12):11-12.
    It is suggested that the impetus to generate models is probably the most fundamental point of connection between mysticism and psychology. In their concern with the relation between ‘unseen’ realms and the ‘seen’, mystical maps parallel cognitive models of the relation between ‘unconscious’ and ‘conscious’ processes. The map or model constitutes an explanation employing terms current within the respective canon. The case of language mysticism is examined to illustrate the premise that cognitive models may benefit from an understanding (...)
    Download  
     
    Export citation  
     
    Bookmark   2 citations  
  45. MECHANICS OF MIND: AN INFRASONIC WAVE MODEL OF HUMAN LANGUAGE ACQUISITION AND COMMUNICATION.Varanasi Ramabraham - 2014 - In Twentieth National Symposium on Ultrasonics (NSU-XX), Department of Physics, Ravenshaw University, cuttack and Ultrasonics Society of India, 24th-25th January, 2014.
    Ideas about human consciousness and mental functions will be analyzed and developed using cognitive science information available in the Upanishads, Brahmajnaana, Advaita and Dvaita schools of thought. -/- The analysis and development so done will be used to theorize and give scheme of human language acquisition and communication process clubbing with Sabdabrahma Siddhanta/Sphota Vaada which put forward infrasonic wave oscillator issuing pulses in infrasonic range and are reflected as brain waves. -/- Thus a brain-wave modulation/demodulation model of human (...)
    Download  
     
    Export citation  
     
    Bookmark  
  46. Complexity Perspectives on Language, Communication and Society.Albert Bastardas-Boada & Àngels Massip-Bonet (eds.) - 2013 - Berlin: Springer.
    The “language-communication-society” triangle defies traditional scientific approaches. Rather, it is a phenomenon that calls for an integration of complex, transdisciplinary perspectives, if we are to make any progress in understanding how it works. The highly diverse agents in play are not merely cognitive and/or cultural, but also emotional and behavioural in their specificity. Indeed, the effort may require building a theoretical and methodological body of knowledge that can effectively convey the characteristic properties of phenomena in human terms. New complexity (...)
    Download  
     
    Export citation  
     
    Bookmark   6 citations  
  47. From Models to Simulations.Franck Varenne - 2018 - London, UK: Routledge.
    This book analyses the impact computerization has had on contemporary science and explains the origins, technical nature and epistemological consequences of the current decisive interplay between technology and science: an intertwining of formalism, computation, data acquisition, data and visualization and how these factors have led to the spread of simulation models since the 1950s. -/- Using historical, comparative and interpretative case studies from a range of disciplines, with a particular emphasis on the case of plant studies, the author shows how (...)
    Download  
     
    Export citation  
     
    Bookmark   15 citations  
  48. You’ve come a long way, baby: the evolution of feminine identity models on the example of contemporary language of advertising.Natalia Anna Michna - 2016 - Polish Journal of Aesthetics 41 (2):99-117.
    The article presents the evolution of the language of advertising from the 1960s to the present, presenting various images of women in advertising. Simultaneously a theoretical analysis has been carried out of the demands of second-wave feminism, which exerted significant influence on the creation of images of women in the mass media. The objective of our comparison of feminist theory with advertising practice is an attempt to answer the question of whether the present media image of women liberated from (...)
    Download  
     
    Export citation  
     
    Bookmark  
  49. Language death and diversity: philosophical and linguistic implications.Lajos L. Brons - 2014 - The Science of Mind 52:243-260.
    This paper presents a simple model to estimate the number of languages that existed throughout history, and considers philosophical and linguistic implications of the findings. The estimated number is 150,000 plus or minus 50,000. Because only few of those remain, and there is no reason to believe that that remainder is a statistically representative sample, we should be very cautious about universalistic claims based on existing linguistic variation.
    Download  
     
    Export citation  
     
    Bookmark   3 citations  
  50. In Pursuit of Unification of Conceptual Models: Sets as Machines.Sabah Al-Fedaghi - manuscript
    Conceptual models as representations of real-world systems are based on diverse techniques in various disciplines but lack a framework that provides multidisciplinary ontological understanding of real-world phenomena. Concurrently, systems’ complexity has intensified, leading to a rise in developing models using different formalisms and diverse representations even within a single domain. Conceptual models have become larger; languages tend to acquire more features, and it is not unusual to use different modeling languages for different components. This diversity has caused problems with consistency (...)
    Download  
     
    Export citation  
     
    Bookmark  
1 — 50 / 950