TS

Tejas Subramaniam

Student of math and economics @ Stanford University
229 karmaJoined Pursuing an undergraduate degree

Bio

Stanford student (math/economics). Formerly intern at Rethink Priorities (animal welfare) and J-PAL South Asia (IDEA Initiative). 

Comments
27

I didn’t say they fell under the ethics of killing, I was using killing as an example of a generic rights violation under a plausible patient-centered deontological theory to illustrate the difference between “a rights violation happening to one person and help coming for a separate person as an offset” and “one’s harm being directly offset.”

(I agree that it seems a bit more unclear if potential people can have rights, even if they can have moral consideration, and in particular rights to not be brought into existence, but I think it’s very plausible.)

Note, however, that I think the question of whether there can be deontic side-constraints regarding our treatment of animals is unclear even conditioning on deontology. Many deontologist philosophers – like Huemer – are uncertain whether animals have “rights” (as a patient-centered deontologist would put it), even though they think (1) humans have rights and (2) animals are still deserving of moral consideration. Deontologists sometimes resort to something like “deontology for people, consequentialism for animals” (although some other deontologists, like Nozick, thought that this was insufficient for animals). 

I think offsetting emissions and offsetting meat consumption are comparable under utilitarianism, but much less comparable under most deontological moral theories, if you think animals have rights. For instance, if you killed someone and donated $5,000 to the Malaria Consortium, that seems worse – from a deontological perspective – than if you just did nothing at all, because the life you kill and the life you save are different people, and many deontological theories are built on the “separateness of persons.” In contrast, if you offset your CO2 emissions, you’re offsetting your effect on warming, so you don’t kill anyone to begin with (because it’s not like your CO2 emissions cause warming that hurts agent A, and then your offset reduces temperatures to benefit agent B). It might be similarly problematic to offset your contribution to air pollution, though, because the effects of air pollution happen near the place where the pollution actually happened. 

Why do you think excruciating pain is 10k as intense as disabling pain? If I use these conversion factors (p. 30) instead, chicken welfare campaigns seem to win. 

Do you think there are promising ways to slow down growth in aquaculture?

Somewhat relevant (takes the hard proves-too-much stance): https://www.econlib.org/archives/2014/10/dear_identity_p.html

She co-authored a piece a few months back about finding AI safety emotionally compelling. I’d be interested in her thoughts on the following two questions related to that!

  • How worried should we be about suspicious convergence between AI safety being one of the most interesting/emotionally compelling questions to think about and it being the most pressing problem? There used to be a lot of discussion around 2015 about how it seemed like people were working on AI safety because it’s really fun and interesting to think about, rather than because it’s actually that pressing. I think that argument is pretty clearly false, but I’d be curious how she views this post as interacting with those concerns. 
  • It seems a bit like the post doesn’t draw a clean distinction between capabilities and safety. I agree that, to some extent, they’re inseparable (the people building transformative AI should care about making it safe), but how does she view the downside risks of, e.g., some of the most compelling parts of AI work being capabilities-related? More generally, how worried should we be, as a community, about how interconnected safety and capabilities work are? 
    • Somewhat related: As Patrick Collison puts it, people working on making more effective engineered viruses aren’t high-status among people working on pandemic prevention, so why are capabilities researchers high-status among safety researchers? 
    • (I have a decent sense of different answers within the community – this is not really a top concern of mine – but I’d nonetheless be interested in her take! My sense is that (1) the distinction isn’t nearly as clean since you want to build AI and make it go safely and (2) it’s good for capabilities work to be more safety-geared than the counterfactual.) 

Thanks for writing this up! I disagree for a few reasons:

  • This feels more like a problem at the point between “alternative proteins have scaled up and we’ve replaced a bunch of meat” and “this results in a meat ban.” It seems possible to me that moral advocacy efforts can happen after alternative proteins have scaled up, but before there are laws to stop factory farming for food entirely. I don’t think alternative proteins replacing, say, 80% of meat will result in people thinking non-meat uses of animals is morally okay in a lock-in kind of way. 
  • I think a lot of people’s moral reasoning about animals is posthoc/based on cognitive dissonance. That is, people like eating meat, or it’s a valuable part of their culture, and their moral intuitions around animal exploitation are built around that. So it seems plausible to me that moral advocacy efforts become substantially more effective if we’re able to quickly replace one of the biggest uses of animals. 
  • I’m not sure I’m compelled by the mechanism for lock-in. One mechanism appears to be overconfidence/complacency as a society, which reduces the drive toward moral progress. This seems somewhat plausible, but it feels like this is possible to solve (for instance, animal advocacy organizations pivot toward other uses of animals, and are more able to dedicate resources focused on animal advocacy). Another mechanism seems to be that “letting automobiles replace horses as practical transport instead of listening to the horse advocates and becoming better humans, humanity has lost a great opportunity to do something for the animals for moral reasons, and do so by accepting an economic loss.” But I guess I’m not sure why – in either the horse case or the factory farming case – this is a unique opportunity. I don’t think the existence of factory farming necessarily strengthens the argument, to an average person, about the urgency of animal advocacy, because if people don’t buy the moral reasoning for caring about animals, I’m not sure the scale of suffering that exists currently affects whether they buy the moral reasoning. So in the case of horses, for example, I don’t think it was easier to convince people that horses matter before they were replaced as practical transport. 
  • I feel like this is just intractable. Meat has the advantage of being embedded in culture and identity for generations. Without proposing any alternative, and going entirely through the moral route, means going up against this generational idea that eating meat is okay. Success seems hard. I’m wary of taking such a risk, when there’s also the possibility of factory farming for food persisting into the future (and I’d guess, in business-as-usual scenarios, it remains a bigger problem than other kinds of factory farming). I will also say I’m not convinced that expanding our moral circle to animals helps expand our moral circle to things like digital minds in the far future, though that’s a conversation for another day. 
  • I’m uncomfortable about this argument for nonconsequentialist reasons. If factory farming is a grave injustice that ought be abolished (even if you’re a consequentialist who buys moral uncertainty), it seems like letting it stay for much longer and taking a huge risk that it stays forever because you want to do it for the right reasons could be a massive negligent injustice in itself. It feels like, in a moral way, saying “it’s bad to hire more beat cops to deter crime, because deterring crime through fear doesn’t convince anyone that their crime is wrong.” One reason a lot of people would find that intuitively bad is because it feels like it’s instrumentalizing the victims of crime for a dubious future consequence. 

These two pieces are of interest. Lots of similar conclusions.

Load more