This is an encyclopedia entry on consequentializing. It explains what consequentializing is, what makes it possible, why someone might be motivated to consequentialize, and how to consequentialize a non-consequentialist theory.
Maximalism is the view that an agent is permitted to perform a certain type of action if and only if she is permitted to perform some instance of this type, where φ-ing is an instance of ψ-ing if and only if φ-ing entails ψ-ing but not vice versa. Now, the aim of this paper is not to defend maximalism, but to defend a certain account of our options that when combined with maximalism results in a theory that accommodates the idea (...) that a moral theory ought to be morally harmonious—that is, ought to be such that the agents who satisfy the theory, whoever and however numerous they may be, are guaranteed to produce the morally best world that they have the option of producing. I argue that, for something to count as an option for an agent, it must, in the relevant sense, be under her control. And I argue that the relevant sort of control is the sort that we exercise over our reasons-responsive attitudes by being both receptive and reactive to reasons. I call this sort of control rational control, and I call the view that φ-ing is an option for a subject if and only if she has rational control over whether she φs rationalism. When we combine this view with maximalism, we get rationalist maximalism, which I argue is a promising moral theory. (shrink)
Blame is multifarious. It can be passionate or dispassionate. It can be expressed or kept private. We blame both the living and the dead. And we blame ourselves as well as others. What’s more, we blame ourselves, not only for our moral failings, but also for our non-moral failings: for our aesthetic bad taste, gustatory self-indulgence, or poor athletic performance. And we blame ourselves both for things over which we exerted agential control (e.g., our voluntary acts) and for things over (...) which we lacked such control (e.g., our desires, beliefs, and intentions). I argue that, despite this manifest diversity in our blaming practices, it’s possible to provide comprehensive account of blame. Indeed, I propose a set of necessary and sufficient conditions that aims to specify blame’s extension in terms of its constitution as opposed to its function. And I argue that this proposal has a number of advantages beyond accounting for blame in all its disparate forms. For one, it can account for the fact that one’s having had control over whether one was to φ is a necessary condition for one’s being fittingly blamed for having φ-ed. For another, it can account for why, unlike fitting shame, fitting blame is always deserved, which in turn explains why there is something morally problematic about ridding oneself of one’s fitting self-blame (e.g., one’s fitting guilt). (shrink)
On what I take to be the standard account of supererogation, an act is supererogatory if and only if it is morally optional and there is more moral reason to perform it than to perform some permissible alternative. And, on this account, an agent has more moral reason to perform one act than to perform another if and only if she morally ought to prefer how things would be if she were to perform the one to how things would be (...) if she were to perform the other. I argue that this account has two serious problems. The first, which I call the latitude problem, is that it has counterintuitive implications in cases where the duty to be exceeded is one that allows for significant latitude in how to comply with it. The second, which I call the transitivity problem, is that it runs afoul of the plausible idea that the one-reason-morally-justifies-acting-against-another relation is transitive. What’s more, I argue that both problems can be overcome by an alternative account, which I call the maximalist account. (shrink)
In this paper, I take it for granted both that there are two types of blameworthiness—accountability blameworthiness and attributability blameworthiness—and that avoidability is necessary only for the former. My task, then, is to explain why avoidability is necessary for accountability blameworthiness but not for attributability blameworthiness. I argue that what explains this is both the fact that these two types of blameworthiness make different sorts of reactive attitudes fitting and that only one of these two types of attitudes requires having (...) been able to refrain from φ-ing in order for them to be fitting. (shrink)
Some right acts have what philosophers call moral worth. A right act has moral worth if and only if its agent deserves credit for having acted rightly in this instance. And I argue that an agent deserves credit for having acted rightly if and only if her act issues from an appropriate set of concerns, where the appropriateness of these concerns is a function what her ultimate moral concerns should be. Two important upshots of the resulting account of moral worth (...) are that (1) an act can have moral worth even if it doesn’t manifest a concern for doing what’s right and that (2) an act can lack moral worth even if it is performed for the right reasons. (shrink)
It seems that we can be directly accountable for our reasons-responsive attitudes—e.g., our beliefs, desires, and intentions. Yet, we rarely, if ever, have volitional control over such attitudes, volitional control being the sort of control that we exert over our intentional actions. This presents a trilemma: (Horn 1) deny that we can be directly accountable for our reasons-responsive attitudes, (Horn 2) deny that φ’s being under our control is necessary for our being directly accountable for φ-ing, or (Horn 3) deny (...) that the relevant sort of control is volitional control. This paper argues that we should take Horn 3. (shrink)
As Socrates famously noted, there is no more important question than how we ought to live. The answer to this question depends on how the reasons that we have for living in various different ways combine and compete. To illustrate, suppose that I've just received a substantial raise. What should I do with the extra money? I have most moral reason to donate it to effective charities but most self-interested reason to spend it on luxuries for myself. So, whether I (...) should live my life as I have most moral reason to live it or as I have most self-interested reason to live it depends on how these and other sorts of reasons combine and compete to determine what I have most reason to do, all things considered. This Element seeks to figure out how different sorts of reasons combine and compete to determine how we ought to live. (shrink)
There is, on a given moral view, an agent-centered restriction against performing acts of a certain type if that view prohibits agents from performing an instance of that act-type even to prevent two or more others from each performing a morally comparable instance of that act-type. The fact that commonsense morality includes many such agent-centered restrictions has been seen by several philosophers as a decisive objection against consequentialism. Despite this, I argue that agent-centered restrictions are more plausibly accommodated within a (...) consequentialist framework than within the more standard side-constraint framework. For I argue that when we combine agent-relative consequentialism with a Kantian theory of value, we arrive at a version of consequentialism, which I call 'Kantsequentialism', that has several advantages over the standard side-constraint approach to accommodating constraints. What’s more, I argue that Kantsequentialism doesn’t have any of the disadvantages that critics of consequentializing have presumed that such a theory must have. (shrink)
IN THIS PAPER, I make a presumptive case for moral rationalism: the view that agents can be morally required to do only what they have decisive reason to do, all things considered. And I argue that this view leads us to reject all traditional versions of actâ€consequentialism. I begin by explaining how moral rationalism leads us to reject utilitarianism.
This paper concerns Warren Quinn’s famous “The Puzzle of the Self-Torturer.” I argue that even if we accept his assumption that practical rationality is purely instrumental such that what he ought to do is simply a function of how the relevant options compare to each other in terms of satisfying his actual preferences that doesn’t mean that every explanation as to why he shouldn’t advance to the next level must appeal to the idea that so advancing would be suboptimal in (...) terms of the satisfaction of his actual preferences. Rather, we can admit that his advancing would always be optimal, but argue that advancing isn’t always what he ought to do given that advancing sometimes fails to meet some necessary condition for being what he ought to do. For instance, something can be what he ought to do only if it’s an option for him. What’s more, something can be what he ought to do only if it’s something that he can do without responding inappropriately to his reasons—or, so, I argue. Thus, the solution to the puzzle is, I argue, to realize that, in certain circumstances, advancing is not what the self-torturer ought to do given that he can do so only by responding inappropriately to his reasons. (shrink)
Maximalism is the view that if an agent is permitted to perform a certain type of action (say, baking), this is in virtue of the fact that she is permitted to perform some instance of this type (say, baking a pie), where φ-ing is an instance of ψ-ing if and only if φ-ing entails ψ-ing but not vice versa. Now, the point of this paper is not to defend maximalism, but to defend a certain account of our options that when (...) combined with maximalism results in a theory that both avoids the sorts of objections that have typically been levelled against maximalism and accommodates the plausible idea that a moral theory must be collectively successful in the sense that everyone’s satisfying the theory guarantees that our theory-given aims will be best achieved. I argue that, for something to count as an option for an agent, it must, in the relevant sense, be under her control. And I argue that the relevant sort of control is the sort that we exercise over our reasons-responsive attitudes (e.g., our beliefs, desires, and intentions) by being both receptive and reactive to reasons. I call this sort of control rational control, and I call the view that φ-ing is an option for an agent if and only if she has rational control over whether she φs rationalism. When we combine this view with maximalism, we get rationalist maximalism, which I argue is a promising moral theory. (shrink)
The performance of one option can entail the performance of another. For instance, baking an apple pie entails baking a pie. Now, suppose that both of these options—baking a pie and baking an apple pie—are permissible. This raises the issue of which, if either, is more fundamental than the other. Is baking a pie permissible because it’s permissible to bake an apple pie? Or is baking an apple pie permissible because it’s permissible to bake a pie? Or are they equally (...) fundamental, as they would be if they were both permissible because, say, they both accord with Kant’s categorical imperative? I defend the view that the permissibility of an option that entails another is more fundamental than the permissibility of the option that it entails. That is, I defend maximalism: the view that if an agent is permitted to perform a certain type of action (say, baking a pie), this is in virtue of the fact that she is permitted to perform some instance of this type (say, baking an apple pie), where φ-ing is an instance of ψ-ing if and only if φ-ing entails ψ-ing but not vice versa. If maximalism is correct, then, as I show, most theories of morality and rationality must be revised. (shrink)
I argue that when determining whether an agent ought to perform an act, we should not hold fixed the fact that she’s going to form certain attitudes (and, here, I’m concerned with only reasons-responsive attitudes such as beliefs, desires, and intentions). For, as I argue, agents have, in the relevant sense, just as much control over which attitudes they form as which acts they perform. This is important because what effect an act will have on the world depends not only (...) on which acts the agent will simultaneously and subsequently perform, but also on which attitudes she will simultaneously and subsequently form. And this all leads me to adopt a new type of practical theory, which I call rational possibilism. On this theory, we first evaluate the entire set of things over which the agent exerts control, where this includes the formation of certain attitudes as well as the performance of certain acts. And, then, we evaluate individual acts as being permissible if and only if, and because, there is such a set that is itself permissible and that includes that act as a proper part. Importantly, this theory has two unusual features. First, it is not exclusively act-orientated, for it requires more from us than just the performance of certain voluntary acts. It requires, in addition, that we involuntarily form certain attitudes. Second, it is attitude-dependent in that it holds that which acts we’re required to perform depends on which attitudes we’re required to form. I then show how these two features can help us both to address certain puzzling cases of rational choice and to understand why most typical practical theories (utilitarianism, virtue ethics, rational egoism, Rossian deontology, etc.) are problematic. (shrink)
The performance of one option can entail the performance of another. For instance, I have the option of baking a pumpkin pie as well as the option of baking a pie, and the former entails the latter. Now, suppose that I have both reason to bake a pie and reason to bake a pumpkin pie. This raises the question: Which, if either, is more fundamental than the other? Do I have reason to bake a pie because I have reason to (...) perform some instance of pie-baking—perhaps, pumpkin-pie baking? Or do I have reason to bake a pumpkin pie because I have reason to bake a pie? Or are they equally fundamental, as they would be if, say, I had reason to do each because each would have optimal consequences? The aim of this paper is to compare two possible answers to this question—omnism and maximalism—and to argue that the latter is preferable. Roughly speaking, maximalism is the view that only those options that are not entailed by any other option are to be assessed in terms of whether they have some feature (such as that of having optimal consequences), whereas omnism is the view that all options are to be assessed in terms of whether they have this feature. I argue that there are at least two reasons to prefer maximalism, for it is able to overcome two critical problems with omnism. (shrink)
In this paper, I argue that we have obligations not only to perform certain actions, but also to have certain attitudes (such as desires, beliefs, and intentions), and this despite the fact that we rarely, if ever, have direct voluntary control over our attitudes. Moreover, I argue that whatever obligations we have with respect to actions derive from our obligations with respect to attitudes. More specifically, I argue that an agent is obligated to perform an action if and only if (...) it’s the action that she would perform if she were to have the attitudes that she ought to have. This view, which I call attitudism, has three important implications. First, it implies that an adequate practical theory must not be exclusively act-orientated. That is, it must require more of us than just the performance of certain voluntary acts. Second, it implies that an adequate practical theory must be attitude-dependent. That is, it must hold that what we ought to do depends on what attitudes we ought to have. Third, it implies that no adequate practical theory can require us to perform acts that we would not perform even if we were to have the attitudes that we ought to have. I then show how these implications can help us both to address certain puzzling cases of rational choice and to understand why most typical practical theories (utilitarianism, rational egoism, virtue ethics, Rossian deontology, etc.) are mistaken. (shrink)
I argue that rule consequentialism sometimes requires us to act in ways that we lack sufficient reason to act. And this presents a dilemma for Parfit. Either Parfit should concede that we should reject rule consequentialism (and, hence, Triple Theory, which implies it) despite the putatively strong reasons that he believes we have for accepting the view or he should deny that morality has the importance he attributes to it. For if morality is such that we sometimes have decisive reason (...) to act wrongly, then what we should be concerned with, practically speaking, is not with the morality of our actions, but with whether our actions are supported by sufficient reasons. We could, then, for all intents and purposes just ignore morality and focus on what we have sufficient reason to do, all things considered. So if my arguments are cogent, they show that Parfit’s Triple Theory is either false or relatively unimportant in that we can, for all intents and purposes, simply ignore its requirements and just do whatever it is that we have sufficient reason to do, all things considered. (shrink)
An act that accords with duty has moral worth if and only if the agent’s reason for performing it is the same as what would have motivated a perfectly virtuous agent to perform it. On one of the two leading accounts of moral worth, an act that accords with duty has moral worth if and only if the agent’s reason for performing it is the fact that it’s obligatory. On the other, an act that accords with duty has moral worth (...) if and only if the agent’s reason for performing it is the fact that it has that feature of obligatory acts that makes them obligatory. I argue that both views are incorrect, providing counterexamples to each. I then argue that, on the correct account, an act can have moral worth only if its agent is motivated out of a fundamental concern for the things that ultimately matter. (shrink)
Imagine both that (1) S1 is deliberating at t about whether or not to x at t' and that (2) although S1’s x-ing at t' would not itself have good consequences, good consequences would ensue if both S1 x's at t' and S2 y's at t", where S1 may or may not be identical to S2 and where t < t' ≤ t". In this paper, I consider how consequentialists should treat S2 and the possibility that S2 will y at (...) t". At one end of the spectrum, consequentialists would hold that, in deciding whether or not to x at t', S1 should always treat S2 as a force of nature over which she has no control and, thus, treat the possibility that S2 will y at t" as she would the possibility that a hurricane will take a certain path. On this view, S1 is to predict whether or not S2 will y and act accordingly. At the other end of the spectrum, consequentialists would hold that S1 should always treat S2 as someone available for mutual cooperation and, thus, treat the possibility that S2 will y at t" as something to be relied upon. On this view, S1 is to rely on S2’s cooperation and so play her part in the best cooperative scheme involving the two of them. A third and intermediate position would be to hold that whether S1 should treat S2 as a force of nature or as someone available for mutual cooperation depends on whether S1 can see to it that S2 will y at t" by, say, having the right set attitudes. I’ll argue for this third position. As we’ll see, an important implication of this view is that consequentialists should be concerned not just with an agent’s voluntary actions but also with their involuntary acquisitions of various mental attitudes, such as beliefs, desires, and intentions. Indeed, I will argue that consequentialists should hold both that (1) an agent’s most fundamental duty is to have all those attitudes that she has decisive reason to have and only those attitudes that she has sufficient reason to have and that (2) she has a derivative duty to perform an act x if and only if her fulfilling this fundamental duty ensures that she x’s. Thus, I argue (as Donald Regan did before me) that consequentialism should not be exclusively act-orientated – that it should require agents not only to perform certain voluntary actions but also to have certain attitudes. In the process, I develop a new version of consequentialism, which I call attitude-consequentialism. (The latest version of this paper can always be found at: https://dl.dropboxusercontent.com/u/14740340/Consequentialism%20and%20Coordination%20Problems.pdf) -/- . (shrink)
The performance of one option can entail the performance of another. For instance, I have the option of baking a pumpkin pie as well as the option of baking a pie, and the former entails the latter. Now, suppose that both of these options are permissible. This raises the issue of which, if either, is more fundamental than the other. Is baking a pie permissible because it’s permissible to perform some instance of pie-baking, such as pumpkin-pie baking? Or is baking (...) a pumpkin pie permissible because it’s permissible to bake a pie? Or are they equally fundamental, as they would be if they were both permissible because, say, they both have optimal consequences? The aim of this paper is to compare two alternative responses to this issue—omnism and maximalism—and to argue that the latter is preferable. Roughly speaking, maximalism is the view that only those options that are not entailed by any other option are to be assessed in terms of whether they have some right-making feature F (such as that of having optimal consequences), whereas omnism is the view that all options are to be assessed in terms of whether they are F. I argue that maximalism is preferable to omnism because it provides a more plausible solution to the problem of act versions and is not subject to any problems of its own. And if I’m right about maximalism’s being preferable to omnism, then most moral theories, which are all versions of omnism, need significant revision. (shrink)
Following Shelly Kagan’s useful terminology, foundational consequentialists are those who hold that the ranking of outcomes is at the foundation of all moral assessment. That is, they hold that moral assessments of right and wrong, virtuous and vicious, morally good and morally bad, etc. are all ultimately a function of how outcomes rank. But foundational consequentialists disagree on what is to be directly evaluated in terms of the ranking of outcomes, which is to say that they disagree on what the (...) primary evaluative focal point is. Act-consequentialists take acts to be the primary evaluative focal point. They evaluate acts in terms of how their outcomes rank (the higher ranked the outcome, the morally better the act), but evaluate everything else in terms of the morally best acts. Thus, the morally best rules are those that would, if internalized, most reliably lead us to perform the morally best acts. Rule-consequentialists, by contrast, take rules to be the primary evaluative focal point. They evaluate rules according to how their outcomes rank and then assess everything else in terms of the morally best rules. Thus, the morally best acts are those that conform to the morally best rules. In this paper, I argue that foundational consequentialists should not take the primary evaluative focal point (or points) to be acts, rules, virtues, or even everything. In so doing, I argue against act-consequentialism, rule-consequentialism, and global consequentialism. But my project is not entirely negative, for I argue that the primary evaluative focal point should be a complex of acts and attitudes. In the end, then, I claim that foundational consequentialists should accept a new kind of consequentialism, which I call attitude-consequentialism. (shrink)
In this dissertation, I argue that commonsense morality is best understood as an agent-relative consequentialist theory, that is, as a theory according to which agents ought always to bring about what is, from their own individual perspective, the best available state of affairs. I argue that the agent-relative consequentialist can provide the most plausible explanation for why it is wrong to commit a rights violation even in order to prevent a number of other agents from committing comparable rights violations: agents (...) bear a special responsibility for their own actions and consequently a state of affairs where an agent has herself committed a rights violation is worse, from her perspective, than a state of affairs where a number of other agents have committed comparable rights violations. I also argue that agent-relative consequentialism can accommodate moral options despite being a theory which requires agents always to do the best they can. (shrink)
Create an account to enable off-campus access through your institution's proxy server.
Monitor this page
Be alerted of all new items appearing on this page. Choose how you want to monitor it:
Email
RSS feed
About us
Lorem ipsum dolor sit amet, consectetur adipisicing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur. Excepteur sint occaecat cupidatat non proident, sunt in culpa qui officia deserunt mollit anim id est laborum.