In G.K. Chesterton’s The Man who was Thursday, six of the seven anarchists named after different days of the week turn out to be secret policemen. Chesterton’s hero Syme finds himself opposed to not just a disparate group of anarchists, but to the unified forces of authority. A similar thing seems to have happened in recent years to JerryFodor. When Fodor published The Language of Thought in 1975 his targets were, as he says, ‘a mixed bag’: (...) reductionists, behaviourists, empiricists, operationalists, holists and various followers of Wittgenstein. But today these disparate targets have become distilled into one movement, which Fodor calls ‘pragmatism’. Fodor sees pragmatism (‘perhaps the worst idea that philosophy ever had’) everywhere, and one aim of this sequel to The Language of Thought is to stamp it out. (shrink)
Contemporary philosophy has had a difficult relationship with its own history. One extreme view conceives of the task of philosophy purely in terms of solving certain given problems, and considers the history of philosophy to have no more relevance to this project than the history of physics has to physics itself. Certainly the history of philosophy is an important intellectual discipline, they argue, but just as physicists do not need to read Newton’s Principia in order to make progress, philosophers do (...) not need to read Aristotle or Kant. At the other extreme are the sceptics who believe that creative philosophy is, for some reason or another it is no longer possible, and that all that can be done is to provide ‘readings’ of the great thinkers of the past. As the late Burton Dreben is supposed to have remarked: ‘junk is junk, but the history of junk is scholarship’. (shrink)
Proponents of non-conceptual content have recruited it for various philosophical jobs. Some epistemologists have suggested that it may play the role of “the given” that Sellars is supposed to have exorcised from philosophy. Some philosophers of mind (e.g., Dretske) have suggested that it plays an important role in the project of naturalizing semantics as a kind of halfway between merely information bearing and possessing conceptual content. Here I will focus on a recent proposal by JerryFodor. In a (...) recent paper he characterizes non-conceptual content in a particular way and argues that it is plausible that it plays an explanatory role in accounting for certain auditory and visual phenomena. So he thinks that there is reason to believe that there is non-conceptual content. On the other hand, Fodor thinks that non-conceptual content has a limited role. It occurs only in the very early stages of perceptual processing prior to conscious awareness. My paper is examines Fodor’s characterization of non-conceptual content and his claims for its explanatory importance. I also discuss if Fodor has made a case for limiting non-conceptual content to non-conscious, sub-personal mental states. (shrink)
Fodor's thinking on modularity has been influential throughout a range of the areas studying cognition, chiefly as a prod for positive work on modularity and domain-specificity. In _The Mind Doesn't Work That Way_, Fodor has developed the dark message of _The Modularity of Mind_ regarding the limits to modularity and computational analyses. This paper offers a critical assessment of Fodor's scepticism with an eye to highlighting some broader issues in play, including the nature of computation and the (...) role of recent empirical developments in the cognitive sciences in assessing Fodor's position. (shrink)
In the third chapter of his book Psychosemantics , Jerry A. Fodor argues that the truth of meaning holism (the thesis that the content of a psychological state is determined by the totality of that state's epistemic liaisons) would be fatal for intentionalistic psychology. This is because holism suggests that no two people are ever in the same intentional state, and so a psychological theory that generalizes over such states will be composed of generalizations which fail to generalize. (...)Fodor then sets out to show that there is no reason to believe in holism by arguing that its primary foundation (i.e. functional-role semantics), when properly understood (i.e. when construed as a two-factor theory of content), is demonstrably false. In this paper, I argue two claims. First, I try to show that Fodor has seriously misrepresented two-factor theories and that his arguments against his strawman do nothing to indicate the falsity of the genuine article. Second, I argue that if one accepts meaning holism in the form of a two-factor theory, there is no particular reason to think that one is hereby committed to the futility of intentionalistic psychology. In making this point, I make a brief excursion into the psychological literature during which I discuss the belief perseverance phenomenon, the encoding specificity hypothesis, and a problem in human deductive reasoning. My second argument leads to a discussion of how such a psychology could be developed even if no two people are ever in the same intentional state. (shrink)
I consider the opening to a paper by JerryFodor referring to graffiti in the subway stations and what Helen Beebee once said about it in her essay writing guide. I used to just pass over that stuff, but now I find it may be more important.
JerryFodor has argued that concept acquisition cannot be a psychological or “rational-causal” process, but can only be a “brute-causal” process of acquisition. This position generates the “doorknob DOORKNOB” problem: why are concepts typically acquired on the basis of experience with items in their extensions? I argue that Fodor’s taxonomy of causal processes needs supplementation, and characterize a third type: what I call “intelligible-causal processes.” Armed with this new category I present what I regard as a (...) better response than Fodor’s to the doorknob DOORKNOB problem. (shrink)
Objections to AI and computational cognitive science are myriad. Accordingly, there are many different reasons for these attacks. But all of them come down to one simple observation: humans seem a lot smarter that computers -- not just smarter as in Einstein was smarter than I, or I am smarter than a chimpanzee, but more like I am smarter than a pencil sharpener. To many, computation seems like the wrong paradigm for studying the mind. (Actually, I think there are deeper (...) and darker reasons why AI, especially, is so often the brunt of polemics, see Dietrich, 2000.) But the truth is this: AI is making exciting progress, and will one day make a robot as intelligent as a person; indeed the robot will be conscious. And all this is because of another truth: the computational paradigm is the best thing to come down the pike since the wheel. (shrink)
In their book What Darwin Got Wrong, JerryFodor and Massimo Piattelli-Palmarini construct an a priori philosophical argument and an empirical biological argument. The biological argument aims to show that natural selection is much less important in the evolutionary process than many biologists maintain. The a priori argument begins with the claim that there cannot be selection for one but not the other of two traits that are perfectly correlated in a population; it concludes that there cannot be (...) an evolutionary theory of adaptation. This article focuses mainly on the a priori argument. (shrink)
Review of the books: Jerry A. Fodor. Concepts: Where Cognitive Science went wrong. Oxford, UK: Oxford University Press, 1998, 174 pp., ISBN 0-19-823636-0. Geoffrey C. Bowker and Susan Leigh Star. Sorting things out: Classification and its consequences. Cambridge, MA: The MIT Press, 1999, 377 pp., ISBN 0-262-02461-6.
This paper examines a recent, influential argument for individualism in psychology defended by JerryFodor and others, what I call the argument from causal powers. I argue that this argument equivocates on the crucial notion of "causal powers", and that this equivocation constitutes a deep problem for arguments of this type. Relational and individualistic taxonomies are incompatible, and it does not seem in general to be possible to factor the former into the latter. The distinction between powers and (...) properties plays a central role in my argument. (shrink)
The Language of Thought program has a suicidal edge. JerryFodor, of all people, has argued that although LOT will likely succeed in explaining modular processes, it will fail to explain the central system, a subsystem in the brain in which information from the different sense modalities is integrated, conscious deliberation occurs, and behavior is planned. A fundamental characteristic of the central system is that it is “informationally unencapsulated” -- its operations can draw from information from any cognitive (...) domain. The domain general nature of the central system is key to human reasoning; our ability to connect apparently unrelated concepts enables the creativity and flexibility of human thought, as does our ability to integrate material across sensory divides. The central system is the holy grail of cognitive science: understanding higher cognitive function is crucial to grasping how humans reach their highest intellectual achievements. But according to Fodor, the founding father of the LOT program and the related Computational Theory of Mind (CTM), the holy grail is out of reach: the central system is likely to be non-computational (Fodor 1983, 2000, 2008). Cognitive scientists working on higher cognitive function should abandon their efforts. Research should be limited to the modules, which for Fodor rest at the sensory periphery (2000).1 Cognitive scientists who work in the symbol processing tradition outside of philosophy would reject this pessimism, but ironically, within philosophy itself, this pessimistic streak has been very influential, most likely because it comes from the most well-known proponent of LOT and CTM. Indeed, pessimism about centrality has become assimilated into the mainstream conception of LOT. (Herein, I refer to a LOT that appeals to pessimism about centrality as the “standard LOT”). I imagine this makes the standard LOT unattractive to those philosophers with a more optimistic approach to what cognitive science can achieve.. (shrink)
In the early part of the paper, I attempt to explain a dispute between two parties who endorse the compositionality of language but disagree about its implications: Paul Horwich, and JerryFodor and Ernest Lepore. In the remainder of the paper, I challenge the thesis on which they are agreed, that compositionality can be taken for granted. I suggest that it is not clear what compositionality involves nor whether it obtains. I consider some kinds of apparent counterexamples, and (...) compositionalist responses to them in terms of covert indexicality and unspecific meanings. I argue that the last option is the best for most of the cases I consider. I conclude by stressing, as against Horwich and Fodor and Lepore, that the appropriate question concerns the extent to which compositionality obtains in a natural language, rather than whether it obtains or not, so that the answer is essentially messy, requiring detailed consideration of a wide range of examples. (shrink)
Although Paul Churchland and JerryFodor both subscribe to the so-called theory-theory– the theory that folk psychology (FP) is an empirical theory of behavior – they disagree strongly about FP’s fate. Churchland contends that FP is a fundamentally flawed view analogous to folk biology, and he argues that recent advances in computational neuroscience and connectionist AI point toward development of a scientifically respectable replacement theory that will give rise to a new common-sense psychology. Fodor, however, wagers that (...) FP will be largely preserved and vindicated by scientific investigations of behavior. Recent findings by developmental psychologists, I argue, will push both Churchlandians and Fodorians toward the pessimistic view that FP is a misguided theory that will never be displaced, because it is, so to speak, built into our cognitive systems. I explore the possibility of preserving optimism by rejecting the theory-theory and adopting the simulation theory, a competing view developed by Robert Gordon, Alvin Goldman, and Jane Heal. According to simulationists, common-sense interpretation of behavior is accomplished by means of pretense-like operations that deploy the cognitive system’s own reasoning capabilities in a disengaged manner. Since on this view no theory-like set of principles would be needed, the simulation theory seems to enjoy a simplicity advantage over the theory-theory. Steven Stich and Shawn Nichols, however, contend that as the cognitive system would require special mechanisms for disengaged operation, the simplicity question cannot be resolved until suitable computational models are developed. I describe a set of models I have constructed to meet this need, and I discuss the contribution such models can make to determining FP’s fate. (shrink)
(1) In the first part of this paper, I review Chomsky's meandering journey from the formalism/mentalism of Syntactic Structures, through several methodological positions, to the minimalist theory of his latest work. Infected with mentalism from first to last, each and every position vitiates Chomsky's repeated claims that his theories will provide useful guidance to later theories in such fields as cognitive psychology and cognitive neuroscience. With the guidance of his insights, he claims, psychologists and neuroscientists will be able to avoid (...) costly dead-end lines of research. -/- (2) This never happened. As I have shown, this never could have happened. (See Johnston, 2018). What has happened, instead, is that current neurolinguistic research (with the arguable exception of the now-dated Lemma Model of Willem Levelt) proceeds without reference to Chomsky. It also wholeheartedly rejects the mentalism of the associated Language of Thought theory of JerryFodor. (See Johnston, 2018). -/- (3) I make this argument in the first part of this paper. I would also like to point out that most of my argument was developed in 1972, when I was a graduate student. I know of no other sustained criticisms of Chomsky at that time, and certainly none along the lines I had developed back then. -/- (4) In the second part of this paper, I present my own account of the methodology of science. When I was a graduate student, philosophy of science was dominated by an attempt to describe a methodology common to all the specific sciences, i.e. Hempel’s deductive-nomological model. These days, Hempel’s emphasis on the methodological unity of science has been rejected by such “dis-unity” philosophers of science as Ian Hacking, Patrick Suppes and Nancy Cartwright (see Cat, 2021). -/- (5) I view this change as the swing of a pendulum or, to change the metaphor, a journey from one end point of a continuum to another. As the level of abstraction at which one tries to describe scientific method is raised, the descriptions become increasingly general. Whether or not unity-of-science theories become so general as to be vacuous, is ultimately a subjective judgment. And so I expect that philosophers will eventually become tired of increasingly specific “close to the workbench” descriptions of how scientists work, and begin to turn back to methodological “big pictures”, finding in them powerful abstractions rather than empty irrelevancies. -/- (6) In the second part of this paper, I present my own account of the methodology of science, which I would situate somewhere between the “unity” and “dis-unity” accounts. However, I am not a scientist. My own views about scientific method have three origins: -/- (6a) my work as a graduate student from 1966 until I passed my comprehensive exams in 1973 (at a different university); -/- (6b) reading every issue of Scientific American from 1972 until nearly 2000, (at which point I continued to read it only sporadically, since I concluded that, around that time, it had evolved from a serious science magazine to a popular science magazine); and -/- (6c) my three-year immersion in the cognitive neuroscience of language after I retired, based on repeated study of and note-taking for (Banesh & Compton, 2018), (Kemmerer, 2015), several other books and, finally, numerous articles not hidden behind a paywall. -/- So, as always with my writings: caveat emptor. (shrink)
[Empiricism, Naturalism, and Ideas] The author analyses the modern reception of key themes in Hume’s philosophy during the past century. The first part presents Hume’s version of three such themes – empiricism, naturalism and the theory of ideas. The following three parts give an exposition of modern forms of each of these themes, with the choice of modern reception being directed to those contemporary authors who not only developed Hume’s motifs in the most original way, but who also explicitly traced (...) the origin of their modern theory to Hume. For this reason, in the second part, which deals with the reception of empiricism in logical positivism, Hans Reichenbach and his treatment of Hume’s problem of inductive knowledge is discussed. In the third part, dealing with naturalism, the obvious choice is the most influential version of this doctrine in the work of W. V. O. Quine. The fourth part deals with the modern reception of Hume’s theory of ideas in a recent monograph by JerryFodor. The author considers Hume’s naturalism as the most live part of Hume’s legacy. Empirismus has, after all, been considerably transformed in content, or has even been rejected by later philosophers; while Fodor’s updating of the theory of ideas does not offer an adequate answer to the question of the place of thinking and intentionality in the material world. (shrink)
The grand opposition between theories of the mind which is presented in this book will be familiar, in its broad outlines, to many readers. On the one side we have the Cartesians, who understand the mind in terms of representation, causation and the inner life; on the other we have the Wittgensteinians, who understand the mind in terms of activity, normativity and its external embedding in its bodily and social environment. In this book—one of a pair, the second of which (...) has yet to be translated—Vincent Descombes puts up a spirited defence of the Wittgensteinian approach. The Cartesian approach, which he calls ‘mental philosophy’, and which is exemplified most typically in the ‘cognitivism’ of JerryFodor, is fundamentally mistaken, he argues, since it underestimates, neglects or ignores both the active and external characteristics of the mind.1 Instead we should2 understand the mind in terms of a human being’s participation in a culture or a ‘form of life’, a form of engagement which is structured by norms rather than causal laws. This ‘anthropological holism’ draws not only upon the work of Wittgenstein, but also on Le´vi-Strauss, Lacan and, among other things, on the role of fiction in shaping our selfunderstanding. (shrink)
Between the end of the nineteenth century and the first twenty years of the twentieth century, stimulated by the impetuous development of logical studies and taking inspiration from Leibniz's idea of a characteristica universalis, the three founding fathers of the analytic tradition in philosophy, i.e., Frege, Russell, and Wittgenstein, started to talk of a logically perfect language, as opposed to natural languages, all feeling that the latter were inadequate to their (different) philosophical purposes. In the second half of the twentieth (...) century, however, the very idea of a logically perfect language ceased for various reasons to seem attractive to analytic philosophers. Thus, it might appear that this idea could be classified together with the many other bizarre ideas that from time to time surface in the history of philosophy-an idea that perhaps had a beneficial impact on the development of twentieth century logic, but which can now be put to rest. In this brief note, I contend that this conclusion may be too hasty. Indeed, if a well-known empirical hypothesis advanced in 1975 by JerryFodor turns out to be true, then there is a logically perfect language, after all. More precisely, I argue that, if it exists, Fodor's language of thought possesses the main characteristics a logically perfect language is required to have. (shrink)
JerryFodor deemed informational encapsulation ‘the essence’ of a system’s modularity and argued that human perceptual processing comprises modular systems, thus construed. Nowadays, his conclusion is widely challenged. Often, this is because experimental work is seen to somehow demonstrate the cognitive penetrability of perceptual processing, where this is assumed to conflict with the informational encapsulation of perceptual systems. Here, I deny the conflict, proposing that cognitive penetration need not have any straightforward bearing on the conjecture that perceptual processing (...) is composed of nothing but informationally encapsulated modules, the conjecture that each and every perceptual computation is performed by an informationally encapsulated module, and the consequences perceptual encapsulation was traditionally expected to have for a perception-cognition border, the epistemology of perception and cognitive science. With these points in view, I propose that particularly plausible cases of cognitive penetration would actually seem to evince the encapsulation of perceptual systems rather than refute/problematize this conjecture. (shrink)
JerryFodor and Ernest Lepore (1999/2002) frame the debate over meaning holism in terms of a distinction between meaning atomism and meaning anatomism. The former holds that the meaning of an expression E is determined by some relation between E and some extra-linguistic entity. The latter holds that the meaning of E is at least partly determined by some of E’s “inward” relations (IRs) with other expressions in the very language. They (1992) argue that meaning anatomism inevitably collapses (...) into meaning holism, which is the view that the meaning of E is determined by E’s IRs with every other expression in the very language because there is no principled distinction for the anatomist to divide the meaning-determining IRs from the non-meaning-determining ones. In response, the non-holistic anatomist urges that Fodor and Lepore’s no-principled-basis consideration is groundless because the lack of a generally accepted criterion for such a distinction does not undermine the viability of the distinction itself. While this point is well taken, I think that Fodor and Lepore are onto an important question here. That is, what does it make non-holistic anatomism distinctive from its holistic counterpart if without a principled basis for the distinction among IRs? I look into this question and give an alternative argument from Fodor and Lepore’s to suggest that non-holistic anatomism cannot bypass the no-principled-basis consideration. The non-holistic anatomist will need a principled distinction in kind between IRs to back her point. (shrink)
JerryFodor, by common agreement, is one of the world’s leading philosophers. At the forefront of the cognitive revolution since the 1960s, his work has determined much of the research agenda in the philosophy of mind and the philosophy of psychology for well over 40 years. This special issue dedicated to his work is intended both as a tribute to Fodor and as a contribution to the fruitful debates that his work has generated. One philosophical thesis that (...) has dominated Fodor’s work since the 1960s is realism about the mental. Are there really mental states, events and processes? From his first book, Psychological Explanation (1968), onwards, Fodor has always answered this question with a resolute yes. From his early rejection of Wittgensteinian and behaviourist conceptions of the mind, to his later disputes with philosophers of mind of the elminativist ilk, he has always been opposed to views that try to explain away mental phenomena. On his view, there are minds, and minds can change the world. (shrink)
One of JerryFodor’s many seminal contributions to philosophy of mind was his inner sentence theory of belief and desire. To believe that p is to have a subpersonal inner sentence in one’s “belief-box” that means that p, and to desire that q is to have a subpersonal inner sentence in one’s “desire-box” that means that q. I will distinguish between two accounts of box-inclusion that exhaust the options: liberal and restrictive. I will show that both accounts have (...) the mistaken implication that in certain cases there can be radical but “secret” changes in a subject’s beliefs and desires. I will suggest that the correct moral to draw is that we should instead accept what Eric Schwitzgebel has called a “surface-level” theory of belief and desire. (shrink)
Maarten Boudry and Jerry Coyne have written a piece, forthcoming in Philosophical Psychology, called “Disbelief in Belief,” in which they criticize my recent paper “Religious credence is not factual belief” (2014, Cognition 133). Here I respond to their criticisms, the thrust of which is that we shouldn’t distinguish religious credence from factual belief, contrary to what I say. I respond that their picture of religious psychology undermines our ability to distinguish common religious people from fanatics. My response will appear (...) in the same issue as their paper. (shrink)
I find it interesting that AI researchers don't use concepts very often in their theorizing. No doubt they feel no pressure to. This is because most AI researchers do use representations which allow a system to chunk up its environment, and basically all we know about concepts is that they are representations which allow a system to chunk up its environment.
In "A Modal Argument for Narrow Content" ("Journal of Philosophy", LXXXVIII, 1991, pp 5-26), JerryFodor proposes a necessary condition for the distinctness of causal powers. He uses this condition to support psychological individualism. I show that Fodor's argument relies on inconsistent interpretations of his condition on distinct causal powers. Moreover, on no consistent interpretation does Fodor's condition yield the results claimed for it.
ABSTRACT. Fodor characterizes concepts as consisting of two dimensions: one is content, which is purely denotational/broad, the other the Mentalese vehicle bearing that content, which Fodor calls the Mode of Presentation (MOP), understood "syntactically." I argue that, so understood, concepts are not interpersonally sharable; so Fodor's own account violates what he calls the Publicity Constraint in his (1998) book. Furthermore, I argue that Fodor's non-semantic, or "syntactic," solution to Frege cases succumbs to the problem of providing (...) interpersonally applicable functional roles for MOPs. This is a serious problem because Fodor himself has argued extensively that if Fregean senses or meanings are understood as functional/conceptual roles, then they can't be public, since, according to Fodor, there are no interpersonally applicable functional roles in the relevant senses. I elaborate on these relevant senses in the paper. (shrink)
This research explores the critical study of language and thought. It is a response to Wittgenstein and Fodor who believe in priority, superiority, and individuality of language and thought. In this paper, we can resolve this language and thought dichotomy to discuss the concept of priority issue. We shall argue that language and thought are modular concepts inside mind and here are also other modules present in our mind. This paper determines the role of language and thought and their (...) priority in mind. Furthermore, this article shows inconsistency in priority, superiority and individuality. It also explores the role of quantification theory which discusses about the structure of thought and language, as well as about the model logic theory which discusses about the semantic part of language an. (shrink)
Radical concept nativism is the thesis that virtually all lexical concepts are innate. Notoriously endorsed by JerryFodor (1975, 1981), radical concept nativism has had few supporters. However, it has proven difficult to say exactly what’s wrong with Fodor’s argument. We show that previous responses are inadequate on a number of grounds. Chief among these is that they typically do not achieve sufficient distance from Fodor’s dialectic, and, as a result, they do not illuminate the central (...) question of how new primitive concepts are acquired. To achieve a fully satisfactory response to Fodor’s argument, one has to juxtapose questions about conceptual content with questions about cognitive development. To this end, we formulate a general schema for thinking about how concepts are acquired and then present a detailed illustration. (shrink)
JerryFodor wrote the following assessment of Danto’s importance in 1993: “Danto has done something I’ve been very much wanting to do: namely, reconsider some hard problems in aesthetics in the light of the past 20 years or so of philosophical work on intentionality and representation” (Fodor 1993, p. 41). Fodor is absolutely right: some of Danto’s work could be thought of as the application of some influential ideas about perception that Fodor also shared. The (...) problem is that these ideas have turned out to be false. Both Danto and Fodor are modularist: they both think that perception is an encapsulated process that is in no way influenced by any kind of non-perceptual processing (see, e.g., Fodor 1983, Pylyshyn 1984). Many of Danto’s famous and influential arguments rely very directly on this modularist assumption. There is now, however, a wealth of evidence against modularism of the strong kind held to by Danto and Fodor. We now know that perceptual experience is not determined entirely by the retinal input: our visual processing is influenced at various point in a top-down manner. What we know and what kinds of visual stimuli we have encountered previously deeply influence how the retinal input is processed. The empirical literature on this is vast and conclusive (for an overview, see Teufel and Nanay 2017 but see also the references in the last section of this paper). What does this mean for Danto’s views on art and perception? While one of Danto’s premises may turn out to be false, the history and examples he gave are valuable and bear repeating. Even more importantly, Danto’s aesthetics can in part be separated out from his modularism, leading us to draw slightly different but arguably even more interesting conclusions from famous thought experiments such as the Gallery of Indiscernibles. (shrink)
In his latest book, The Elm and the Expert (1994), Fodor notoriously rejects the notion of narrow content as superfluous. He envisions a scientific intentional psychology that adverts only to broad content properties in its explanations. I argue that Fodor's change in view is only apparent and that his previous position (1985-1991) is extensionally equivalent to his "new" position (1994). I show that, despite what he says narrow content is for in his (1994), Fodor himself has previously (...) never appealed to the notion of narrow content in explaining Frege cases and cases involving the so-called deferential concepts. And for good reason: his notion of narrow content (1985-91) couldn't explain them. The only apparent change concerns his treatment of Twin Earth cases. However, I argue that the notion of broad content that his purely informational semantics delivers is, in some interesting sense, equivalent to the mapping notion of narrow content he officially gave up. For his pure informational semantics fails to avoid assigning disjunctive content to twins, since nomic covariations take care not only actual but also counterfactual contexts into account. I show that none of the attempts made by Fodor to block this consequence of his theory works. The present notion of broad content he now operates with is therefore in a position to take over all the important jobs that his previous notion of narrow content could do. (shrink)
In The Mind Doesn’t Work that Way, JerryFodor argues that mental representations have context sensitive features relevant to cognition, and that, therefore, the Classical Computational Theory of Mind (CTM) is mistaken. We call this the Globality Argument. This is an in principle argument against CTM. We argue that it is self-defeating. We consider an alternative argument constructed from materials in the discussion, which avoids the pitfalls of the official argument. We argue that it is also unsound and (...) that, while it is an empirical issue whether context sensitive features of mental representations are relevant to cognition, it is empirically implausible. (shrink)
In this paper I want to propose an argument to support JerryFodor’s thesis (Fodor 1983) that input systems are modular and thus informationally encapsulated. The argument starts with the suggestion that there is a “grounding problem” in perception, i. e. that there is a problem in explaining how perception that can yield a visual experience is possible, how sensation can become meaningful perception of something for the subject. Given that visual experience is actually possible, this invites (...) a transcendental argument that explains the conditions of its possibility. I propose that one of these conditions is the existence of a visual module in Fodor’s sense that allows the step from sensation to object-identifying perception, thus enabling visual experience. It seems to follow that there is informationally encapsulated nonconceptual content in visual perception. (shrink)
hroughout medieval Christianity, religious works of art emerged to illustrate the teachings of the Bible for the largely illiterate population. What, then, is the significance of the psychoactive mushrooms hiding in plain sight in the artwork and icons of many European and Middle-Eastern churches? Does Christianity have a psychedelic history? -/- Providing stunning visual evidence from their anthropological journey throughout Europe and the Middle East, including visits to Roslyn Chapel and Chartres Cathedral, authors Julie and Jerry Brown document the (...) role of visionary plants in Christianity. They retrace the pioneering research of R. Gordon Wasson, the famous “sacred mushroom seeker,” on psychedelics in ancient Greece and India, and among the present-day reindeer herders of Siberia and the Mazatecs of Mexico. Challenging Wasson’s legacy, the authors reveal his secret relationship with the Vatican that led to Wasson’s refusal to pursue his hallucinogen theory into the hallowed halls of Christianity. -/- Examining the Bible and the Gnostic Gospels, the authors provide scriptural support to show that sacred mushrooms were the inspiration for Jesus’ revelation of the Kingdom of Heaven and that he was initiated into these mystical practices in Egypt during the Missing Years. They contend that the Trees of Knowledge and of Immortality in Eden were sacred mushrooms. -/- Uncovering the role played by visionary plants in the origins of Judeo-Christianity, the authors invite us to rethink what we know about the life of Jesus and to consider a controversial theory that challenges us to explore these sacred pathways to the divine. (shrink)
The purpose of this paper is to present the essence of Ayn Rand's theory of rational egoism and to indicate how it is the only ethical theory that can provide a foundation for ethics in business. Justice, however, cannot be done to the breadth and depth of Rand's theory in so short a space as this article; consequently, I have provided the reader with a large number of references for further study. At minimum, Ayn Rand's theory, because of its originality (...) and challenge to establishment theories, should be included in all business ethics courses and discussions of business ethics. (shrink)
In the TACITUS project for using commonsense knowledge in the understanding of texts about mechanical devices and their failures, we have been developing various commonsense theories that are needed to mediate between the way we talk about the behavior of such devices and causal models of their operation. Of central importance in this effort is the axiomatization of what might be called commonsense metaphysics. This includes a number of areas that figure in virtually every domain of discourse, such as granularity, (...) scales, time, space, material, physical objects, shape, causality, functionality, and force. Our effort has been to construct core theories of each of these areas, and then to define, or at least characterize, a large number of lexical items in terms provided by the core theories. In this paper we discuss our methodological principles and describe the key ideas in the various domains we are investigating. (shrink)
Fodor argued that learning a concept by hypothesis testing would involve an impossible circularity. I show that Fodor's argument implicitly relies on the assumption that actually φ-ing entails an ability to φ. But this assumption is false in cases of φ-ing by luck, and just such luck is involved in testing hypotheses with the kinds of generative random sampling methods that many cognitive scientists take our minds to use. Concepts thus can be learned by hypothesis testing without circularity, (...) and it is plausible that this is how humans in fact acquire at least some of their concepts. (shrink)
Abstract: It is the aim of this paper to establish a temporal and cultural link between entheogen-use1 in Classical mystery cults and their possible use in a segment of the early Christian Gnostic Church. As early Christianity was heavily influenced by the Classical world in which it first developed, it is essential to examine the evidence of entheogen-use within Classical mystery cults, and explore their possible influence on the development of Christian ritual. We will first present textual evidence from the (...) New Testament and Nag Hammadi Library (NHL) that indicate possible use of ritual entheogens. Next we will examine entheogen-use by some of the main religious sects of the ancient Mediterranean world as part of their liturgical practice between c. fifth century BCE and the third century CE. Then we will show that the different schools of philosophical thought examined were quite fluid, and that there was considerable overlap among schools with respect to several key philosophies and practices. Ultimately the subject of entheogen-use in the Classical world necessarily falls into the realm of plausible speculation mainly due to the lack of primary sources. However, should the entheogenic hypothesis2 prove to be correct, it promises to be quite useful in future research and in the translation of original documents. (shrink)
One puzzling feature of Plato’s Republic is the First City or ‘city of pigs’. Socrates praises the First City as a “true”, “healthy” city, yet Plato abandons it with little explanation. I argue that the problem is not a political failing, as most previous readings have proposed: the First City is a viable political arrangement, where one can live a deeply Socratic lifestyle. But the First City has a psychological corollary, that the soul is simple rather than tripartite. Plato sees (...) this ‘First Soul’ as an inaccurate model of moral psychology, and so rejects it, along with its political analogue. (shrink)
In light of new historical evidence regarding ethnomycologist R. Gordon Wasson’s correspondence with art historian Erwin Panofsky, this article provides an in-depth analysis of the presence of entheogenic mushroom images in Christian art within the context of the controversy between Wasson and philologist John Marco Allegro over the identification of a Garden of Eden fresco in the 12th century Chapel of Plaincourault in France. It reveals a compelling financial motive for Wasson’s refusal to acknowledge that this fresco represents Amanita muscaria, (...) as well as for Wasson’s reluctance to pursue his hypothesis regarding the entheogenic origins of religion into Christian art and artifacts. While Wasson’s view – that the presence of psychoactive mushrooms in the Near and Middle East ended around 1000 BCE – prevailed and stymied research on entheogens in Christianity for decades, a new generation of 21st century researchers has documented growing evidence of A. muscaria and psilocybin-containing mushrooms in Christian art, consistent with ethnobotanist Giorgio Samorini’s typology of mushroom trees. This article presents original photographs, taken during fieldwork at churches and cathedrals throughout Europe and the Middle East, that confirm the presence of entheogenic mushrooms in Christian art: in frescoes, illuminated manuscripts, mosaics, sculptures, and stained glass windows. Based on this iconic evidence, the article proposes a psychedelic gospels theory and addresses critiques of this theory by art historians, ardent advocates, medieval historians, and conservative Catholics. It calls for the establishment of an Interdisciplinary Committee on the Psychedelic Gospels to independently evaluate the growing body of evidence of entheogenic mushrooms in Christian art in order to resolve a controversial question regarding the possible role of entheogens in the history and origins of Christianity. (shrink)
This paper proposes that Fodor has, or had, a personal reason to avoid the creative writing department, to do with his opposition to inferential role semantics.
Highlights: -/- • Breastfeeding and breastfeeding support can contribute to mitigating climate change. • Achieving global nutrition targets will save more emissions than fuel-switching. • Breastfeeding support programmes support a just transition. • This work can support the expansion of mitigation options in energy system models. -/- Abstract: -/- Renewable gas has been proposed as a solution to decarbonise industrial processes, specifically heat demand. As part of this effort, the breast-milk substitutes industry is proposing to use renewable gas as a (...) substitute for fossil natural gas. However, decarbonising the industrial processing of breast-milk substitutes can increase social license for these products, potentially undermining breastfeeding. World Health Organisation nutrition targets aim to increase exclusive breastfeeding to at least 50% globally by 2025 to improve maternal, infant, and young child health and nutrition. This target will have implications for the energy transition. A weakness of existing energy models is that demands for end-use products such as breast-milk substitutes are typically not considered explicitly. This paper develops an analytical framework for explicitly representing infant feeding methods in energy systems models. We compare the emissions saved in Ireland from decarbonising the industrial processing of breast-milk substitutes with renewable gas with the emissions saved by an increase in exclusive breastfeeding to 50% in both Ireland and a key export market, China. We demonstrate that the emissions saved from achieving the minimum global breastfeeding target are greater than when renewable gas is used to displace natural gas in the production of breast-milk substitutes in Ireland. We discuss the decarbonisation of breast-milk substitutes in relation to the principle of justice as non-maleficence, a principle based on the commitment to avoid harm, a novel application of a principle of justice. We conclude that breastfeeding support can be considered a demand-side measure for mitigating climate change by reducing the demand for energy services to produce breast-milk substitutes. A key recommendation is to position breastfeeding support as both a public health and a climate justice issue that is relevant for a just transition. The framework developed for this paper could be applied to support the inclusion of a wider range of mitigation options with social justice outcomes in energy system models. [Open access]. (shrink)
Create an account to enable off-campus access through your institution's proxy server.
Monitor this page
Be alerted of all new items appearing on this page. Choose how you want to monitor it:
Email
RSS feed
About us
Lorem ipsum dolor sit amet, consectetur adipisicing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur. Excepteur sint occaecat cupidatat non proident, sunt in culpa qui officia deserunt mollit anim id est laborum.