Purpose Prescription stimulant use as “cognitive enhancers” has been described among undergraduate college students. However, the use of prescription stimulants among future health care professionals is not well characterized. This study was designed to determine the prevalence of prescription stimulant misuse among students at an academic health sciences center. -/- Method Electronic surveys were e-mailed to 621 medical, pharmacy, and respiratory therapy students at East Tennessee State University for four consecutive weeks in fall 2011. Completing the survey was voluntary and (...) anonymous. Surveys asked about reasons for, frequency of, and side effects of nonprescription misuse of prescription stimulants. Given the sensitive material, an opportunity to win one of ten $50 gift cards was used as an incentive. -/- Results Three hundred seventy-two (59.9%) students completed the survey from three disciplines (47.6% medical, 70.5% pharmacy, and 57.6% respiratory therapy). Overall, 11.3% of responders admitted to misusing prescription stimulants. There was more misuse by respiratory therapy students, although this was not statistically significant (10.9% medicine, 9.7% pharmacy, 26.3% respiratory therapy; P = .087). Reasons for prescription stimulant misuse included to enhance alertness/ energy (65.9%), to improve academic performance (56.7%), to experiment (18.2%), and to use recreationally/get high (4.5%). -/- Conclusions Prescription stimulant misuse was prevalent among participating students, but further research is needed to describe prevalence among future health care workers more generally. The implications and consequences of such misuse require further study across professions with emphasis on investigating issues of academic dishonesty (e.g., “cognitive enhancement”), educational quality, and patient safety or health care quality. (shrink)
I present an account of deterministic chance which builds upon the physico-mathematical approach to theorizing about deterministic chance known as 'the method of arbitrary functions'. This approach promisingly yields deterministic probabilities which align with what we take the chances to be---it tells us that there is approximately a 1/2 probability of a spun roulette wheel stopping on black, and approximately a 1/2 probability of a flipped coin landing heads up---but it requires some probabilistic materials to work with. I contend that (...) the right probabilistic materials are found in reasonable initial credence distributions. I note that, with some normative assumptions, the resulting account entails that deterministic chances obey a variant of Lewis's 'principal principle'. I additionally argue that deterministic chances, so understood, are capable of explaining long-run frequencies. (shrink)
Is logic normative for reasoning? In the wake of work by Gilbert Harman and John MacFarlane, this question has been reduced to: are there any adequate bridge principles which link logical facts to normative constraints on reasoning? Hitherto, defenders of the normativity of logic have exclusively focussed on identifying adequate validity bridge principles: principles linking validity facts—facts of the form 'gamma entails phi'—to normative constraints on reasoning. This paper argues for two claims. First, for the time being at least, Harman’s (...) challenge cannot be surmounted by articulating validity bridge principles. Second, Harman’s challenge can be met by articulating invalidity bridge principles: principles linking invalidity facts of the form 'gamma does not entail phi' to normative constraints on reasoning. In doing so, I provide a novel defence of the normativity of logic. (shrink)
The majority of our linguistic exchanges, such as everyday conversations, are divided into turns; one party usually talks at a time, with only relatively rare occurrences of brief overlaps in which there are two simultaneous speakers. Moreover, conversational turn-taking tends to be very fast. We typically start producing our responses before the previous turn has finished, i.e., before we are confronted with the full content of our interlocutor’s utterance. This raises interesting questions about the nature of linguistic understanding. Philosophical theories (...) typically focus on linguistic understanding characterized either as an ability to grasp the contents of utterances in a given language or as outputs of this ability—mental states of one type or another. In this paper, I supplement these theories by developing an account of the process of understanding. I argue that it enables us to capture the dynamic and temporal aspect of understanding and reconcile philosophical investigations with empirical research on language comprehension. (shrink)
The language of phenomenology includes terms such as intentionality, phenom- enon, insight, analysis, sense, not to mention the key term of Edmund Husserl’s manifesto, “the things themselves” to return to . But what does the “things them- selves” properly mean? How come the term is replaced by the “findings” over time? And what are the findings for? The investigation begins by looking at the tricky legacy of the modern turn, trying to clarify ties to past masters, including Francis- co Suárez (...) and Augustine of Hippo . The former, because his influence goes beyond René Descartes reaching undoubtedly Franz Brentano and his students, as well as Martin Heidegger . The latter, because Augustine gives a personal component to the Greek inheritance, marked by the “inward turn .” However, it would not be possible to review the history of thought without the help offered by Jan Patočka's analyses . Patočka discloses the “care” of the Greek philosophers, Plato and Dem- ocritus among others, “for the soul”, we would say with Patočka for “being,” whose sense “does not leave us indifferent” as the leitmotiv of Ancient Philosophy . Nev- ertheless, in his lectures on Plato and Europe, Patočka points out that you must be careful not to confuse the phenomena of things, of existens, with the phenomena of being . Finally, Patočka’s legacy is found in the efforts to reconcile the life-feeling with the modern construction of reality, which means “a radical reconstruction of the naive and natural world of common sense .” In some ways, intentionality is to be revised . (shrink)
Multiple-choice questions have an undeserved reputation for only being able to test student recall of basic facts. In fact, well-crafted mechanically gradable questions can measure very sophisticated cognitive skills, including those engaged at the highest level of Benjamin Bloom’s taxonomy of outcomes. In this article, I argue that multiple-choice questions should be a part of the diversified assessment portfolio for most philosophy courses. I present three arguments broadly related to fairness. First, multiple-choice questions allow one to consolidate subjective decision making (...) in a way that makes it easier to manage. Second, multiple-choice questions contribute to the diversity of an evaluation portfolio by balancing out problems with writing-based assessments. Third, by increasing the diversity of evaluations, multiple-choice questions increase the inclusiveness of the course. In the course of this argument I provide examples of multiple-choice questions that measure sophisticated learning and advice for how to write good multiple-choice questions. (shrink)
Se presenta la traducción de los capítulos 1 y 2 del libro Spinoza and Time del filósofo judío Samuel Alexander, el que deriva de la Cuarta Conferencia en Memoria de Arthur Davis, dictada ante la Jewish Historical Society de Inglaterra, el domingo 1 de mayo, 1921/23 de Nisan, 5681. La traducción responde a la necesidad de contar con un acercamiento en castellano al corpus alexandriano, ya que no existe al día de hoy una traducción total de sus libros. A su (...) vez el traductor encuentra motivación en el redescubrimiento de autores judíos que aborden el tema de la temporalidad. (shrink)
We investigated, experimentally, the contention that the folk view, or naïve theory, of time, amongst the population we investigated is dynamical. We found that amongst that population, ~ 70% have an extant theory of time that is more similar to a dynamical than a non-dynamical theory, and ~ 70% of those who deploy a naïve theory of time deploy a naïve theory that is more similar to a dynamical than a non-dynamical theory. Interestingly, while we found stable results across our (...) two experiments regarding the percentage of participants that have a dynamical or non-dynamical extant theory of time, we did not find such stability regarding which particular dynamical or non-dynamical theory of time they take to be most similar to our world. This suggests that there might be two extant theories in the population—a broadly dynamical one and a broadly non-dynamical one—but that those theories are sufficiently incomplete that participants do not stably choose the same dynamical theory as being most similar to our world. This suggests that while appeals to the ordinary view of time may do some work in the context of adjudicating disputes between dynamists and non-dynamists, they likely cannot do any such work adjudicating disputes between particular brands of dynamism. (shrink)
Philosophers have long noted, and empirical psychology has lately confirmed, that most people are “biased toward the future”: we prefer to have positive experiences in the future, and negative experiences in the past. At least two explanations have been offered for this bias: belief in temporal passage and the practical irrelevance of the past resulting from our inability to influence past events. We set out to test the latter explanation. In a large survey, we find that participants exhibit significantly less (...) future bias when asked to consider scenarios where they can affect their own past experiences. This supports the “practical irrelevance” explanation of future bias. It also suggests that future bias is not an inflexible preference hardwired by evolution, but results from a more general disposition to “accept the things we cannot change”. However, participants still exhibited substantial future bias in scenarios in which they could affect the past, leaving room for complementary explanations. Beyond the main finding, our results also indicate that future bias is stake-sensitive and that participants endorse the normative correctness of their future-biased preferences and choices. In combination, these results shed light on philosophical debates over the rationality of future bias, suggesting that it may be a rational response to empirical realities rather than a brute, arational disposition. (shrink)
In the early twentieth century, many philosophers in America thought that time should be taken seriously in one way or another. George P. Adams (1882-1961) argued that the past, present and future are all real but only the present is actual. I call this theory ‘actualist eternalism’. In this paper, I articulate his novel brand of eternalism as one piece of his metaphysical system and I explain how he argued for the view in light of the best explanations of temporal (...) experience and the present. I argue that his exploitation of analogies between time and modality offer some lessons for current debates about time such as the importance of providing a temporal epistemology. I also extract what I call the temporal boundary problem and argue that it gives rise to an unaddressed challenge for presentists and growing block theorists. (shrink)
The causal efficacy of a material system is usually thought to be produced by the law-like actions and interactions of its constituents. Here, a specific system is constructed and explained that produces a cause that cannot be understood in this way, but instead has novel and autonomous efficacy. The construction establishes a proof-of-feasibility of strong emergence. The system works by utilizing randomness in a targeted and cyclical way, and by relying on sustained evolution by natural selection. It is not vulnerable (...) to standard arguments against strong emergence, in particular ones that assume that the physical realm is causally closed. Moreover, it does not suffer from epiphenomenalism or causal overdetermination. The system uses only standard material components and processes, and is fully consistent with naturalism. It is discussed whether the emergent cause can still be viewed as ‘material’ in the way that term is commonly understood. (shrink)
One of the central figures of philosophy of language- John Langshaw Austin, attributes principles of causation to the mere pragmatic language. Conversely, Kant tried to construct a “free human act” which is independent from any physical determination except its innate motivations via his well-known the phenomenal / noumenal distinction. That kind of Kantian metaphysical ground which addresses to the noumenal field, he obviously tries to establish this behavioral causation again by denying Austinian style pragmatic propositions or illocutionary acts. I claimed (...) that sort of duality between Austin and Kant, creates an epistemological problem with how propositions and actions relate. From a Kantian position, it (indetermination) is overlooked by Austin's propositional doctrine, without being grounded on any universal principle, but only with propositions that embraced by speech act theory. (shrink)
Epistemology is often a problem for libertarianism. Many libertarian texts assume that they need to do more than explain and defend the libertarian conjecture. Instead, they try to offer epistemological support for it (whether empirically or morally); which falsificationism and, more broadly, critical rationalism explains is not possible. Moreover, they often mistake this attempt at support for an explanation of libertarianism (which ought to include an abstract theory of liberty and how it relates to liberty in practice). Therefore, when a (...) criticism of falsificationism appears on a libertarian website it seems useful to reply to it (even though no discussion of libertarianism itself is involved). (shrink)
What it would take to vindicate folk temporal error theory? This question is significant against a backdrop of new views in quantum gravity—so-called timeless physical theories—that claim to eliminate time by eliminating a one-dimensional substructure of ordered temporal instants. Ought we to conclude that if these views are correct, nothing satisfies the folk concept of time and hence that folk temporal error theory is true? In light of evidence we gathered, we argue that physical theories that entirely eliminate an ordered (...) substructure vindicate folk temporal error theory. (shrink)
The authors argue that their target is orthogonal to the realism and instrumentalist debate. I argue that it is born directly from it. While the distinction is helpful in illuminating how some ontological commitments demand a theory of implementation, it's less clear whether different views cleanly map onto the epistemic and metaphysical uses defined in the paper.
The two purposes of this essay. The general philosophical problem with most versions of social libertarianism and how this essay will proceed. The specific problem with liberty explained by a thought-experiment. The positive and abstract theory of interpersonal liberty-in-itself as ‘the absence of interpersonal initiated constraints on want-satisfaction’, for short ‘no initiated impositions’. The individualistic liberty-maximisation theory solves the problems of clashes, defences, and rectifications without entailing interpersonal utility comparisons or libertarian consequentialism. The practical implications of instantiating liberty: three rules (...) of liberty-in-practice, 1) initial ultimate control of one’s body, 2) initial ultimate control of one’s used resources, 3) consensual interpersonal interactions and resource transfers. These rules are economically efficient and maximise general want-satisfaction. Private property and legal remedies are additional practical institutional aspects, but to which ‘initiated impositions’ then apply prima facie. Libertarian law is often mistaken for complete libertarianism. Moral explanations are a separate issue. The three main moral theories imply libertarianism, but it can be morally posited independently of them. Critical rationalism and its application. No empirical or argumentative support for theories. An important ambiguity with ‘justification’. How the epistemology applies to the theory of liberty and its application but remains separable in principle. Conclusion: there are further published explanations but this should be enough to generate useful criticism. Appendix replying to some typical comments. (shrink)
In How We Get Along, philosopher David Velleman compares our social interactions to the interactions among improvisational actors on stage. He argues that we play ourselves - not artificially but authentically, by doing what would make sense coming from us as we really are. And, like improvisational actors, we deal with one another in dual capacities: both as characters within the social drama and as players contributing to the shared performance. In this conception of social intercourse, Velleman finds rational grounds (...) for morality, though not a rational guarantee. He maps a middle course between skepticism and rationalism, arguing that practical reasoning is 'pro-moral' without requiring moral action. The result is what he calls a 'Kinda Kantian metaethics'. How We Get Along is the summation of Velleman's thinking to date, incorporating and unifying previous work on agency, the self, the emotions, narrative and Kantian moral theory. (shrink)
Trust is a topic of longstanding philosophical interest. It is indispensable to every kind of coordinated human activity, from sport to scientific research. Even more, trust is necessary for the successful dissemination of knowledge, and by extension, for nearly any form of practical deliberation and planning. Without trust, we could achieve few of our goals and would know very little. Despite trust’s fundamental importance in human life, there is substantial philosophical disagreement about what trust is, and further, how trusting is (...) normatively constrained and best theorized about in relation to other things we value. This entry is divided into three sections, which explore key (and sometimes interconnected) ethical and epistemological themes in the philosophy of trust: (1) The Nature of Trust; (2) The Normativity of Trust, and (3) The Value of Trust. (shrink)
On Kratzer’s canonical account, modal expressions (like “might” and “must”) are represented semantically as quantifiers over possibilities. Such expressions are themselves neutral; they make a single contribution to determining the propositions expressed across a wide range of uses. What modulates the modality of the proposition expressed—as bouletic, epistemic, deontic, etc.—is context.2 This ain’t the canon for nothing. Its power lies in its ability to figure in a simple and highly unified explanation of a fairly wide range of language use. Recently, (...) though, the canon’s neat story has come under attack. The challenge cases involve the epistemic use of a modal sentence for which no single resolution of the contextual parameter appears capable of accommodating all our intuitions.3 According to these revisionaries, such cases show that the canonical story needs to be amended in some way that makes multiple bodies of information relevant to the assessment of such statements. Here I show that how the right canonical, flexibly contextualist account of modals can accommodate the full range of challenge cases. The key will be to extend Kratzer’s formal semantic account with an account of how context selects values for a modal’s.. (shrink)
Decisions are made under uncertainty when there are distinct outcomes of a given action, and one is uncertain to which the act will lead. Decisions are made under indeterminacy when there are distinct outcomes of a given action, and it is indeterminate to which the act will lead. This paper develops a theory of (synchronic and diachronic) decision-making under indeterminacy that portrays the rational response to such situations as inconstant. Rational agents have to capriciously and randomly choose how to resolve (...) the indeterminacy relevant to a given choice-situation, but such capricious choices once made constrain how they will choose in the future. The account is illustrated by the case of self-interested action in situations where it is indeterminate whether you yourself will survive to benefit or suffer the consequences. The conclusion emphasizes some distinctive anti-hedging predictions of the account. (shrink)
Readers of “Falsificationism Redux” (the rejoinder) may have found it to be another waffling non-explanation of induction and the alleged falsity of falsificationism—or even self-refuting, as its title indicates (redux: brought back, revived, restored). However, it seems worth another round of replies if only because the arguments are fairly typical of the would-be ‘inductivist’ and it might help some people who have yet to see how these arguments fail.
There are advantages to thrift over honest toil. If we can make do without numbers we avoid challenging questions over the metaphysics and epistemology of such entities; and we have a good idea, I think, of what a nominalistic metaphysics should look like. But minimizing ontology brings its own problems; for it seems to lead to error theory— saying that large swathes of common-sense and best science are false. Should recherche philosophical arguments really convince us to give all this up? (...) Such Moorean considerations are explicitly part of the motivation for the recent resurgence of structured metaphysics, which allow a minimal (perhaps nominalistic) fundamental ontology, while avoiding error-theory by adopting a permissive stance towards ontology that can be argued to be grounded in the fundamental. This paper evaluates the Moorean arguments, identifying key epistemological assumptions. On the assumption that Moorean arguments can be used to rule out error-theory, I examine deflationary ‘representationalist’ rivals to the structured metaphysics reaction. Quinean paraphrase, fictionalist claims about syntax and semantics are considered and criticized. In the final section, a ‘direct’ deflationary strategy is outlined and the theoretical obligations that it faces are articulated. The position advocated may have us talking a lot like a friend of structured metaphysics—but with a very different conception of what we’re up to. (shrink)
As it is known, there is no rule satisfying Additivity in the complete domain of bankruptcy problems. This paper proposes a notion of partial Additivity in this context, to be called µ-additivity. We find that µ-additivity, together with two quite compelling axioms, anonymity and continuity, identify the Minimal Overlap rule, introduced by Neill (1982).
What considerations place genuine constraints on an adequate semantics for normative and evaluative expressions? Linguists recognize facts about ordinary uses of such expressions and competent speakers’ judgments about which uses are appropriate. The contemporary literature reflects the widespread assumption that linguists don’t rely upon an additional source of data—competent speakers’ judgments about possible disagreement with hypothetical speech communities. We have several good reasons to think that such judgments are not probative for semantic theorizing. Therefore, we should accord these judgments no (...) probative value for the development of a semantics for our moral terms. Such judgments can no longer be presumed to put pressure on theories according to which our moral expressions share a semantics with ordinary, descriptive terms. Many rivals to pure, Descriptivist theories count among their advantages the ability to accommodate these judgments. If these judgments have no probative value, such theories lose an important source of support. (shrink)
This is an attempt to clarify the nature of extreme, or complete, “wokeness” in its modern sense. The central thesis is that it is an inverted form of fascism, and thereby even worse than some of its critics assume. In fact, it is far worse than ordinary fascism whether or not it is correct to see it as an inverted form. As this is a thesis, it is not a definition. Therefore, this thesis could certainly be mistaken. But if it (...) is, then sound criticisms of the explanation would be needed to refute it. (shrink)
As a philosopher of action, I might be expected to believe that the will is a good thing. Actually, I believe that the will is a great thing - awesome, in fact. But I'm not thereby committed to its being something good. When I say that the will is awesome, I mean literally that it is a proper object of awe, a response that restrains us from abusing the will and moves us rather to use it respectfully, in a way (...) that does it justice. To say that the will is a good thing, however, would imply that having a will is better than not having one, or that using it is better than not using it - neither of which I am prepared to assert as a general rule. Speaking metaphorically, I would say that the will is like a magic wand. In fairy tales, the character who looks upon a magic wand as an unalloyed good is destined to be sadder but wiser in the end. Being a magician isn't better than being an ordinary human, just different; and a magician must value his powers by respecting them and therefore using them appropriately, even sparingly, not by using them as much as possible. (shrink)
While there seems to be much evidence that perceptual states can occur without being conscious, some theorists recently express scepticism about unconscious perception. We explore here two kinds of such scepticism: Megan Peters and Hakwan Lau's experimental work regarding the well-known problem of the criterion -- which seems to show that many purported instances of unconscious perception go unreported but are weakly conscious -- and Ian Phillips' theoretical consideration, which he calls the 'problem of attribution' -- the worry that many (...) purported examples of unconscious perception are not perceptual, but rather merely informational and subpersonal. We argue that these concerns do not undermine the evidence for unconscious perception and that this sceptical approach results in a dilemma for the sceptic, who must either deny that there is unconscious mentality generally or explain why perceptual states are unique in the mind such that they cannot occur unconsciously. Both options, we argue, are problematic. (shrink)
This paper examines narrativism’s claim that the historical past cannot be known once and for all because it must be continuously re-described from the standpoint of the present. We argue that this claim is based on a non sequitur. We take narrativism’s claim that the past must be re-described continuously from the perspective of the present to be the result of the following train of thought: 1) “all knowledge is conceptually mediated”; 2) “the conceptual framework through which knowledge of reality (...) is mediated changes with every new generation of historians”; therefore (narrativism’s claim) “the historical past changes with every new generation of historians”. The idea of an unchanging past, for the narrativist, requires denying premise 1 (all knowledge is conceptually mediated”) and therefore rests on a problematic commitment to the chimerical notion of the past as it is in-itself, wie es eigentlich gewesen. We argue that the narrativist’s conclusion does not follow unless one adds a further premise, namely 3) “it is not possible to view reality through the categorial framework of historical agents”. If one asserts the possibility of grasping reality through the categorial framework of others, be they contemporary or past agents (as much philosophy of history written in an idealist key does), then one no longer has to accept the narrativist’s inference that since the past cannot be known in-itself or independently of conceptual mediation, then it cannot be known as it always was for the historical agents. Narrativism’s inference that the past cannot be known as it always was does not follow from premises 1 and 2 unless one smuggles in another problematic premise, premise 3. In this paper we defend the claim that the past can be known as it always was (not as it is in-itself) by invoking a different conception of the role of conceptual mediation in historical knowledge, one which assumes the possibility of viewing reality through the categorial framework of others. This notion of the role of conceptual mediation in historical knowledge is prevalent in the idealist tradition but, in the interest of brevity, we will defend this notion of mediacy by specific reference to the idealist philosophy of R.G. Collingwood. (shrink)
Inspired by Rudolf Carnap's Der Logische Aufbau Der Welt, David J. Chalmers argues that the world can be constructed from a few basic elements. He develops a scrutability thesis saying that all truths about the world can be derived from basic truths and ideal reasoning. This thesis leads to many philosophical consequences: a broadly Fregean approach to meaning, an internalist approach to the contents of thought, and a reply to W. V. Quine's arguments against the analytic and the a priori. (...) Chalmers also uses scrutability to analyze the unity of science, to defend a conceptual approach to metaphysics, and to mount a structuralist response to skepticism. Based on the 2010 John Locke lectures, Constructing the World opens up debate on central philosophical issues involving language, consciousness, knowledge, and reality. This major work by a leading philosopher will appeal to philosophers in all areas. This entry contains uncorrected proofs of front matter, chapter 1, and first excursus. (shrink)
A series of papers on different aspects of practical knowledge by Roderick Chisholm, Rudolf Haller, J. C. Nyiri, Eva Picardi, Joachim Schulte Roger Scruton, Barry Smith and Johan Wrede.
Corcoran, J. 2005. Counterexamples and proexamples. Bulletin of Symbolic Logic 11(2005) 460. -/- John Corcoran, Counterexamples and Proexamples. Philosophy, University at Buffalo, Buffalo, NY 14260-4150 E-mail: [email protected] Every perfect number that is not even is a counterexample for the universal proposition that every perfect number is even. Conversely, every counterexample for the proposition “every perfect number is even” is a perfect number that is not even. Every perfect number that is odd is a proexample for the existential proposition that some (...) perfect number is odd. Conversely, every proexample for the proposition “some perfect number is odd” is a perfect number that is odd. As trivial these remarks may seem, they can not be taken for granted, even in mathematical and logical texts designed to introduce their respective subjects. One well-reviewed book on counterexamples in analysis says that in order to demonstrate that a universal proposition is false it is necessary and sufficient to construct a counterexample. It is easy to see that it is not necessary to construct a counterexample to demonstrate that the proposition “every true proposition is known to be true” is false–necessity fails. Moreover the mere construction of an object that happens to be a counterexample does not by itself demonstrate that it is a counterexample–sufficiency fails. In order to demonstrate that a universal proposition is false it is neither necessary nor sufficient to construct a counterexample. Likewise, of course, in order to demonstrate that an existential proposition is true it is neither necessary nor sufficient to construct a proexample. This article defines the above relational concepts of counterexample and of proexample, it discusses their surprising history and philosophy, it gives many examples of uses of these and related concepts in the literature and it discusses some of the many errors that have been made as a result of overlooking the challenging subtlety of the proper use of these two basic and indispensable concepts. (shrink)
The philosophical significance of attitudinal autonomy—viz., the autonomy of attitudes such as beliefs—is widely discussed in the literature on moral responsibility and free will. Within this literature, a key debate centres around the following question: is the kind of attitudinal autonomy that’s relevant to moral responsibility at a given time determined entirely by a subject’s present mental structure at that time? Internalists say ‘yes’, externalists say ’no’. In this essay, I motivate a kind of distinctly epistemic attitudinal autonomy, attitudinal autonomy (...) that is relevant to knowledge. I argue that regardless of whether we are externalists or internalists about the kind of attitudinal autonomy that is relevant for moral responsibility, we should be externalists about the kind of autonomy that a belief must have to qualify as knowledge. (shrink)
A new way to transpose the virtue epistemologist’s ‘knowledge = apt belief’ template to the collective level, as a thesis about group knowledge, is developed. In particular, it is shown how specifically judgmental belief can be realised at the collective level in a way that is structurally analogous, on a telic theory of epistemic normativity (e.g., Sosa 2020), to how it is realised at the individual level—viz., through a (collective) intentional attempt to get it right aptly (whether p) by alethically (...) affirming that p. An advantage of the proposal developed is that it is shown to be compatible with competing views—viz., joint acceptance accounts and social-distributive accounts—of how group members must interact in order to materially realise a group belief. I conclude by showing how the proposed judgment-focused collective (telic) virtue epistemology has important advantages over a rival version of collective virtue epistemology defended in recent work by Jesper Kallestrup (2016). (shrink)
A norm of local expert deference says that your credence in an arbitrary proposition A, given that the expert's probability for A is n, should be n. A norm of global expert deference says that your credence in A, given that the expert's entire probability function is E, should be E(A). Gaifman (1988) taught us that these two norms are not equivalent. Here, I provide characterisation theorems which tell us precisely when the norms give different advice. They tell us that, (...) in a good sense, Gaifman's example is the only case where the two norms differ. I suggest that the lesson of the theorems is that Bayesian epistemologists need not concern themselves with the differences between these two kinds of norms. While they are not strictly speaking equivalent, they are equivalent for all philosophical purposes. (shrink)
This paper sets out to defend human genetic engineering with a new bioethical approach, post-humanism, combined with a radical democratic political framework. Arguments for the restriction of human genetic engineering, and specifically germ-line enhancement, are reviewed. Arguments are divided into those which are fundamental matters of faith, or "bio-Luddite" arguments, and those which can be addressed through public policy, or "gene-angst" arguments.The four bio-Luddite concerns addressed are: Medicine Makes People Sick; There are Sacred Limits of the Natural Order; Technologies Always (...) Serve Ruling Interests; The Genome is Too Complicated to Engineer. I argue that these are matters of faith that one either accepts or rejects, and that I reject.The non-fundamentalist or pragmatic concerns I discuss are: Fascist Applications; The Value of Genetic Diversity; The Geneticization of Life; Genetic Discrimination and Confidentiality; Systematically Bad Decisions by Parents; Discrimination Against the Disabled; Unequal Access; The Decline of Social Solidarity. I conclude that all these concerns can be adequately addressed through a proactive regulative framework administered by a liberal democratic state. Therefore, even germ-line genetic enhancement should eventually made available since the potential benefits greatly outweigh the potential risks. (shrink)
The human brain consists of approximately 100 billion electrically active neurones that generate an endogenous electromagnetic field, whose role in neuronal computing has not been fully examined. The source, magnitude and likely influence of the brain's endogenous em field are here considered. An estimate of the strength and magnitude of the brain's em field is gained from theoretical considerations, brain scanning and microelectrode data. An estimate of the likely influence of the brain's em field is gained from theoretical principles and (...) considerations of the experimental effects of external em fields on neurone firing both in vitro and in vivo. Synchronous firing of distributed neurones phase-locks induced em field fluctuations to increase their magnitude and influence. Synchronous firing has previously been demonstrated to correlate with awareness and perception, indicating that perturbations to the brain's em field also correlate with awareness. The brain's em field represents an integrated electromagnetic field representation of distributed neuronal information and has dynamics that closely map to those expected for a correlate of consciousness. I propose that the brain's em information field is the physical substrate of conscious awareness - the cemi field - and make a number of predictions that follow from this proposal. Experimental evidence pertinent to these predictions is examined and shown to be entirely consistent with the cemi field theory. This theory provides solutions to many of the intractable problems of consciousness - such as the binding problem - and provides new insights into the role of consciousness, the meaning of free will and the nature of qualia. It thus places consciousness within a secure physical framework and provides a route towards constructing an artificial consciousness. (shrink)
Since 1964, through 7 revisions, the World Medical Association’s Declaration of Helsinki has stood as an important statement regarding the ethical principles guiding medical research with human participants. It is consulted by ethics review committees, funders, researchers, and research participants. It has been incorporated into national legislation and is routinely invoked to ascertain the ethical appropriateness of clinical trials. There is much to praise about the revision process and the latest revision, which coincides with the declaration’s 50th anniversary. Nevertheless, the (...) proposed declaration contains persistent flaws... (shrink)
Philosophy of mind and cognitive science (e.g., Clark and Chalmers 1998; Clark 2010; Palermos 2014) have recently become increasingly receptive tothe hypothesis of extended cognition, according to which external artifacts such as our laptops and smartphones can—under appropriate circumstances—feature as material realisers of a person’s cognitive processes. We argue that, to the extent that the hypothesis of extended cognition is correct, our legal and ethical theorising and practice must be updated, by broadening our conception of personal assault so as to (...) include intentional harm towards gadgets that have been appropriately integrated. We next situate the theoretical case for extended personal assault within the context of some recent ethical and legal cases and close with some critical discussion. (shrink)
In the April 2002 edition of JCS I outlined the conscious electromagnetic information field theory, claiming that consciousness is that component of the brain's electromagnetic field that is downloaded to motor neurons and is thereby capable of communicating its informational content to the outside world. In this paper I demonstrate that the theory is robust to criticisms. I further explore implications of the theory particularly as regards the relationship between electromagnetic fields, information, the phenomenology of consciousness and the meaning of (...) free will. Using cemi field theory I propose a working hypothesis that shows, among other things, that awareness and information represent the same phenomenon viewed from different reference frames. (shrink)
In Lehrer’s case of the superstitious lawyer, a lawyer possesses conclusive evidence for his client’s innocence, and he appreciates that the evidence is conclusive, but the evidence is causally inert with respect to his belief in his client’s innocence. This case has divided epistemologists ever since Lehrer originally proposed it in his argument against causal analyses of knowledge. Some have taken the claim that the lawyer bases his belief on the evidence as a data point for our theories to accommodate, (...) while others have denied that the lawyer has knowledge, or that he bases his belief on the evidence. In this paper, we move the dialectic forward by way of arguing that the superstitious lawyer genuinely infers his client’s innocence from the evidence. To show that the lawyer’s inference is genuine, we argue in defense of a version of a doxastic construal of the ‘taking’ condition on inference. We also provide a pared-down superstitious lawyer-style case, which displays the key features of the original case without including its complicated and distracting features. But interestingly, although we argue that the lawyer’s belief is based on his good evidence, and is also plausibly doxastically justified, we do not argue that the lawyer knows that his client is innocent. (shrink)
Create an account to enable off-campus access through your institution's proxy server.
Monitor this page
Be alerted of all new items appearing on this page. Choose how you want to monitor it:
Email
RSS feed
About us
Lorem ipsum dolor sit amet, consectetur adipisicing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur. Excepteur sint occaecat cupidatat non proident, sunt in culpa qui officia deserunt mollit anim id est laborum.