In general, I've seen philosophers "bucket" normative ethics into 3 primary categories:

  • Virtue Ethics: Emphasize moral character
  • Deontology: Emphasizes duties or rules
  • Consequentialism: Emphasizes the consequences of actions

See here or here.

However, I would prefer to combine virtue ethics and deontology to create a binary distinction:

  • Means: Virtue Ethics and Deontology
  • Ends: Consequentialism

I'm not an expert philosopher by any means (heh), but this makes more intuitive sense to me. When we think about "how to do good", the 1st clear question is "are you thinking about your actions (means) or the outcomes of those actions (ends)?" For me, virtue ethics and deontology are two ways to think about your actions. i.e. Deontology—Do your actions align with some duty/rule? Or Virtue Ethics—Do your actions align with some moral character traits?

Questions:

  1. Is it actually true that philosophers (generally) give the 3-category version over the 2-category version?
  2. What am I missing about virtue ethics/deontology that implies I shouldn't categorize them both into "means"?
  3. Whatever the answers to #1 and #2, what do you find to be the most helpful categorical breakdown of normative ethics?

Thanks!

10

0
0

Reactions

0
0
Comments7
Sorted by Click to highlight new comments since:

Can I ask why you actually want to catagorize ethics like this at all? I know that it is traditional and it can be very helpful when teaching ethics to set things out like this as if you don’t then students often miss the profound difference between different ethical theories. However a lot of exciting work has never fallen into these catagories, which in any case are basically post hoc classifications of the work of Aristotle, Plato and Mill. Hume’s work for instance is pretty clearly ‘none of the above’ and a lot of good creative work has been done over the past century or more in trying to break down the barriers between these schools (Sidgwick and Parfit being the two biggest names here, but by no means the only ones). Personally I think that there are a lot of good tools and poweful arguments to be found across the ethical spectrum and that so long as you apreciate the true breadty of diversity among ethical theories then breaking them down like this is no longer much help for anything really.

From an EA perspective I think that the one distinction that may be worth paying attention to, and that can fit into your ‘consiquentialism’ Vs ‘deontology and virtue ethics’ distinction, though it is not a perfect fit, are moral theories that can be incorporated into an ‘expected moral value’ framework and those that can’t. This is an important distinction because it places a limit on how far one can go in making precise judgements about what one ought to do in the face of uncertainty which is something that may be of concern to all EAs. However this is a distinction that emerges from practice rather than being baked into moral theories at the level of first principles and there are other aproaches, such as the ‘parliamentary model’ for handelling moral uncertainty, that seek to overcome it.

I think it's a mischaracterisation to think of virtue ethics in terms of choosing the most virtuous actions (in fact, one common objection to virtue ethics is that it doesn't help very much in choosing actions). I think virtue ethics is probably more about being the most virtuous, and making decisions for virtuous reasons. There's a difference: e.g. you're probably not virtuous if you choose normally-virtuous actions for the wrong reasons.

For similar reasons, I disagree with cole_haus that virtue ethicists choose actions to produce the most virtuous outcomes (although there is at least one school of virtue ethics which seems vaguely consequentialist, the eudaimonists. See https://plato.stanford.edu/entries/ethics-virtue). Note however that I haven't actually looked into virtue ethics in much detail.

Edit: contractarianism is a fourth approach which doesn't fit neatly into either division

I think there's a certain prima facie plausibility to the traditional tripartite division. If you just think about the world in general, each of actors, actions, and states seem salient. It wouldn't take much to convince me that--appropriately defined--actors, actions, and states are mutually exclusive and collectively exhaustive in some metaphysical sense.

Once you accept the actors, actions, states division, it makes sense to have ethical theories revolving around each. These corresponds to virtue ethics, deontology and consequentialism.

I think you could fairly convincingly bucket virtue ethics in 'Ends' if you wanted to adopt this schema. A virtue ethicist could be someone who chooses the action that produces the best outcome in terms of personal virtue. They are (sort of) a utilitarian that optimizes for virtue rather than utility and restricts their attention to only themselves rather than the whole world.

Honestly, the consequentialist/virtue/deontological division is not a complete categorization, just a list of the three most common theories; other moral theories simply fall outside of it. The Philpapers survey includes "other" as a fourth option to make a true categorization.

I haven't seen any philosophers use a means/ends binary categorization. But others are often used instead of the consequentialist/virtue/deontological division. For instance, there is the agent-neutral/agent-relative distinction, and there is the thick concept centralism/thin concept centralism distinction. All can be more or less appropriate in different contexts. Overall, philosophers don't worry that much about this question of categorization, so don't worry too much about what they use.

What am I missing about virtue ethics/deontology that implies I shouldn't categorize them both into "means"?

I am not well read on virtue ethics, but it seems to me that both cole haus and richard ngo are right: some kinds of virtue ethics are about the end of achieving a certain personal character, and others are about having actions that align with the right principles.

Whatever the answers to #1 and #2, what do you find to be the most helpful categorical breakdown of normative ethics?

In most contexts I would tend towards stressing something Parfit talks about, between traditionalism (though I forget the exact word he uses) and revisionism. The former being moral theories that try to preserve our ordinary folk moral beliefs, and the latter being moral theories that try to revise them to meet more rigorous principles. Because I think that a relatively high amount of moral disagreement boils down to this. But again there are multiple ways of making categories here, and there is not one universal best option, they are just useful in different ways in different contexts.

What am I missing about virtue ethics/deontology that implies I shouldn't categorize them both into "means"?

I find this grouping weird as I think of virtue ethics (which, full disclosure, is most like the way I intuitively approach applied ethics) as the only one that's really about means rather than ends because it's about what generates your observable behaviors rather than the observable behaviors themselves.

To see what I mean, a deontological ethical system might say "don't kill other humans" so you don't kill other humans based on that rule. In a certain sense this is a means to an end that is similar to what a consequentialist ethics might select for—humans end up not being killed, but because you followed the rule—but it is also an end in itself—you don't do any killing. The only difference between deontological systems and consequentialist systems in this case is that a consequentialist system places the emphasis on (or, if you like, optimizes on) the global outcome while the deontological system focuses more directly on ones observable behavior. It's a question of where you put the pressure in this sense: either on the observable, desired outcomes in the world that imply the actions that generate the desired outcomes (consequentialism) or on the observable behaviors that generate the desired outcomes (deontology).

Virtue ethics is to me the odd-man-out here because it's focused on behavior that is not observable, or at least not observable by anyone other than the subject who is experiencing it. You can observe the outcomes of one who has virtue, which is intended to generate behaviors that will cause desirable outcomes in the world, but it pushes back the focus a level so that you optimize/emphasize the spontaneous generation of normative behavior rather than rules that generate normative behavior or normative behavior to produces a particular outcome.

Is it actually true that philosophers (generally) give the 3-category version over the 2-category version?

Philosophers use lots of different categorization methods for different reasons. The thing to keep in mind is that any groupings are always serving some purpose, so talking in terms of virtue, deon, and consequences is simply one useful way approaching ethical systems that helps us understand them. Other categorizations are possible and frequently used, and there is frequent disagreement among philosophers over where to draw particular lines or whether this or that ethical notion really belongs in some particular category.

Personally I find it helpful to think of these categorizations found throughout philosophy as useful for certain purposes in academic philosophy but not necessary useful to the objectives of philosophy itself except inasmuch as you are addressing your readers and helping them understand an idea in the context of other ideas they are familiar with. It's tempting to get caught up in constructing detailed models of philosophical ideas at the expense of the philosophical act itself.

Whatever the answers to #1 and #2, what do you find to be the most helpful categorical breakdown of normative ethics?

All that being said, I tend to think about descriptive vs. prescriptive approaches. As Hume argued, we can't directly derive the normative from what is normal without metaphysical speculation, and for myself I find no reason to believe it is necessary to speculate in a way that ends up implying any particular norms. Therefore I tend to think about categorizing ethics in terms of whether we are trying to discuss norms in terms of what we find (descriptive) or what we wish we found (prescriptive), my own inclinations running towards the descriptive.

[anonymous]-2
0
0

I don't think the tripartite division is particularly helpful. It smacks of parochialism. It's only been the standard way of breaking down 'normative ethics' among a small clique of analytic philosophers in the Anglophone world - i.e. a few thousand people - beginning sometime in the twentieth century. It's a shame that it has become the default pedagogical tool for introducing students to ethics. It has some merit as such, but students end up thinking that it's 'the' division of ethics, and it invariably ends up occluding more than it illuminates.

If you try and fit most 'canonical' figures in the history of social and political thought into the tripartite division - e.g. Thucydides, Epictetus, Augustine, Montaigne, Voltaire, Rousseau, Marx, Nietzsche, Dewey, Wittgenstein - it becomes immediately apparent that it's an incredibly crude and misleading way of looking at ethics, and assumes a great deal about what 'ethics' is. Let alone if you go beyond the canon and look at more marginal figures, or ethnography/anthropology/cultural history for that matter. As someone else said, most thinkers are sui generis; it is almost always unhelpful to impose these kind of blunt ex post categories on them. The subject is infinitely richer and more complicated than that.

Curated and popular this week
 ·  · 25m read
 · 
Epistemic status: This post — the result of a loosely timeboxed ~2-day sprint[1] — is more like “research notes with rough takes” than “report with solid answers.” You should interpret the things we say as best guesses, and not give them much more weight than that. Summary There’s been some discussion of what “transformative AI may arrive soon” might mean for animal advocates. After a very shallow review, we’ve tentatively concluded that radical changes to the animal welfare (AW) field are not yet warranted. In particular: * Some ideas in this space seem fairly promising, but in the “maybe a researcher should look into this” stage, rather than “shovel-ready” * We’re skeptical of the case for most speculative “TAI<>AW” projects * We think the most common version of this argument underrates how radically weird post-“transformative”-AI worlds would be, and how much this harms our ability to predict the longer-run effects of interventions available to us today. Without specific reasons to believe that an intervention is especially robust,[2] we think it’s best to discount its expected value to ~zero. Here’s a brief overview of our (tentative!) actionable takes on this question[3]: ✅ Some things we recommend❌ Some things we don’t recommend * Dedicating some amount of (ongoing) attention to the possibility of “AW lock ins”[4]  * Pursuing other exploratory research on what transformative AI might mean for animals & how to help (we’re unconvinced by most existing proposals, but many of these ideas have received <1 month of research effort from everyone in the space combined — it would be unsurprising if even just a few months of effort turned up better ideas) * Investing in highly “flexible” capacity for advancing animal interests in AI-transformed worlds * Trying to use AI for near-term animal welfare work, and fundraising from donors who have invested in AI * Heavily discounting “normal” interventions that take 10+ years to help animals * “Rowing” on na
 ·  · 3m read
 · 
About the program Hi! We’re Chana and Aric, from the new 80,000 Hours video program. For over a decade, 80,000 Hours has been talking about the world’s most pressing problems in newsletters, articles and many extremely lengthy podcasts. But today’s world calls for video, so we’ve started a video program[1], and we’re so excited to tell you about it! 80,000 Hours is launching AI in Context, a new YouTube channel hosted by Aric Floyd. Together with associated Instagram and TikTok accounts, the channel will aim to inform, entertain, and energize with a mix of long and shortform videos about the risks of transformative AI, and what people can do about them. [Chana has also been experimenting with making shortform videos, which you can check out here; we’re still deciding on what form her content creation will take] We hope to bring our own personalities and perspectives on these issues, alongside humor, earnestness, and nuance. We want to help people make sense of the world we're in and think about what role they might play in the upcoming years of potentially rapid change. Our first long-form video For our first long-form video, we decided to explore AI Futures Project’s AI 2027 scenario (which has been widely discussed on the Forum). It combines quantitative forecasting and storytelling to depict a possible future that might include human extinction, or in a better outcome, “merely” an unprecedented concentration of power. Why? We wanted to start our new channel with a compelling story that viewers can sink their teeth into, and that a wide audience would have reason to watch, even if they don’t yet know who we are or trust our viewpoints yet. (We think a video about “Why AI might pose an existential risk”, for example, might depend more on pre-existing trust to succeed.) We also saw this as an opportunity to tell the world about the ideas and people that have for years been anticipating the progress and dangers of AI (that’s many of you!), and invite the br
 ·  · 3m read
 · 
Hi all, This is a one time cross-post from my substack. If you like it, you can subscribe to the substack at tobiasleenaert.substack.com. Thanks Gaslit by humanity After twenty-five years in the animal liberation movement, I’m still looking for ways to make people see. I’ve given countless talks, co-founded organizations, written numerous articles and cited hundreds of statistics to thousands of people. And yet, most days, I know none of this will do what I hope: open their eyes to the immensity of animal suffering. Sometimes I feel obsessed with finding the ultimate way to make people understand and care. This obsession is about stopping the horror, but it’s also about something else, something harder to put into words: sometimes the suffering feels so enormous that I start doubting my own perception - especially because others don’t seem to see it. It’s as if I am being gaslit by humanity, with its quiet, constant suggestion that I must be overreacting, because no one else seems alarmed. “I must be mad” Some quotes from the book The Lives of Animals, by South African writer and Nobel laureate J.M. Coetzee, may help illustrate this feeling. In his novella, Coetzee speaks through a female vegetarian protagonist named Elisabeth Costello. We see her wrestle with questions of suffering, guilt and responsibility. At one point, Elisabeth makes the following internal observation about her family’s consumption of animal products: “I seem to move around perfectly easily among people, to have perfectly normal relations with them. Is it possible, I ask myself, that all of them are participants in a crime of stupefying proportions? Am I fantasizing it all? I must be mad!” Elisabeth wonders: can something be a crime if billions are participating in it? She goes back and forth on this. On the one hand she can’t not see what she is seeing: “Yet every day I see the evidences. The very people I suspect produce the evidence, exhibit it, offer it to me. Corpses. Fragments of