by EJT
4 min read 5

27

[Saying an old thing in a new way]

The United States Department of Transportation will pay $11.8 million to save a life. You know what that means? It means that if you come to the United States Department of Transportation with a plan (barriers around the Grand Canyon, wider lanes on the expressway, no left-turns on Sundays, etc. etc. etc.), the United States Department of Transportation will take your plan – snatch the blueprints right out of your hand – and go away and calculate two numbers. The first is the cost: the cold hard cash required to make your plan a reality. The second is the expected number of Americans saved by your plan. They’ll think of every number of Americans that your plan could save and couple it up with the chance that your plan in fact saves that number of Americans. They’ll have each of these pairs be fruitful and multiply and then they’ll sum. I exaggerate and simplify, but only slightly.

Now the United States Department of Transportation – or the US Department of Transportation, as I’m sometimes bold enough to call them – has your two numbers: the cost (measured in dollars) and the benefit (measured in Americans). You’re summoned back into the head honcho’s office and she punches these numbers into the ceremonial calculator: dollars divided by American lives. If the answer is $11.8 million per life or less, she shakes your hand and writes a cheque. If the answer is more than $11.8 million per life, she says ‘Take a hike. Never talk to me again.’ I exaggerate and simplify, but only slightly.

Now. I am an American citizen; I care only about the two inevitables. Do you know what I mean by ‘the inevitables’?

                                                                                                                                            Dash? Elastigirl?

No, no, no. The inevitables: death and taxes. I am an American citizen; I care only about death and taxes. I exaggerate and simplify, but only slightly.

Now every time the head honcho at the US DoT signs a cheque, my taxes go up. That’s bad. But once the DoT project is complete, my chance of getting smeared across America’s highways goes down. That’s good. I’m a typical American. Every new $20 million DoT project is another 6¢ out of my pocket and at least another 0.000000005 ticks down on the old highway-death-o-meter. It’s a good deal. I’m happy to take it. I’ll have what I’m having. I exaggerate and simplify, but only slightly.

But then I come across The Precipice: a book written by one Toby Ord of Oxford University, England. In The Precipice, Ord gives us his best guess of the chance that everything goes wrong this century: 1-in-6, Russian roulette. And ‘everything’ here means everything. We’re talking disaster-movie-without-the-happy-ending. Nuclear conflagration, a modern plague, AI takeover. Armageddon, Contagion, War Games, The Day After Tomorrow, Terminator. Humanity in ruins and remaining so, for all the aeons until our great universe dissipates into fuzz.

But forget that last part. I never went in for that ‘deep time’ stuff anyway. Today’s troubles are enough for today, and my troubles are enough for me. Let’s focus on the here-and-now. This century. My life and the life of my children. Russian roulette. Do I feel lucky? I’m a confident guy. I know my way around a screwdriver. I like to think I’d comport myself pretty well in a total breakdown of the social order. But if things go Dr Strangelove, I’ve just got to come to terms with the fact that one of those ICBMs might have my name on it and that no degree of screwdriving-prowess is going to scratch it off. No man is an island. That bell will toll for me.

No. If it’s to be less than 1-in-6 that I perish from this earth, I need the government. Only they can make a real dent in my chances of getting fried, or of spluttering out my last in some makeshift hospital ward. And not only can the government do it, the price is right. Carl Shulman and Elliott Thornley estimate that $32 billion per year would increase the chance that we make it through the decade by 1-in-200. That’s a small number, but the ceremonial calculator reveals that it’s an even better deal than the one I’m getting from the US DoT: $100 out of my pocket and a big tick down on the old annihilation-o-meter.

1-in-6, 1-in-200. The numbers are unsettling, but they’re also unsettled. Odds are at home in Vegas where they’re born from the motions of dice, cards, and roulette wheels, and can be confirmed by long painful experience. But Ord has never sat and tallied infernos, never watched his supply of civilizations grow and fall and dwindle to nil. The 6s and 200s are estimates, surmises, guesses, and I’ve bought enough used cars in my time to be wary of numbers plucked from thin air. But still. The question remains. What should we do? Should we renew the nuclear treaties or let them lapse? Should we hire scientists to watch for cooked-up superdiseases or save our money? Should we pump the brakes on AI or let it race ahead? If we do nothing, we’ll be guessing that the risk is low. If we do something, we’ll be guessing that the risk is not so low. Whatever we choose, we’ll be making a guess, so we might as well use our best one. Your best guess might differ from Ord’s, but you won’t know what it is until you make it your best: until you sit down with head clear and heart still and look at what we know. Nuclear weapons are younger than your grandparents and they’ve fallen out of planes many, many more times. Rabies kills all its untreated victims, COVID-19 infected almost everyone, and lab-engineered diseases could do both. AI systems are already so big that their creators can’t understand them and so erratic that billion-dollar companies can’t control them, and they’re about to get hundreds of times bigger.

It might feel wrong to say a number. But remember: guessing is not optional. Every action is a choice, and every choice is a guess. So look at what we know; guess, guess, guess; and use your best. If the bombs start falling, how many of us die? If it’s an engineered pandemic, how many of us die? What can we do? How much would it cost? How much would it help? Then get out the calculator. Less than $11.8 million per American and it would save a whole load of non-Americans too? It’s a good deal. Let’s take it.

27

0
0

Reactions

0
0

More posts like this

Comments5


Sorted by Click to highlight new comments since:

The US DOT may technically have official guidance for the value of a statistical life, but I don’t think this actually informs much of the department’s priorities. Most DOT spending is on projects that can be expected to increase the number of roadway deaths due to increased speed and increased miles driven.

The US DOT budget is set by Congress and Congress has allocated most of that budget to programs other than safety (mainly highway expansions). I don’t think DOT is even allowed to enforce its VSL when making grants to states.

The reality is that there are some roadway interventions that are way better than $10MM per life. Protected bike lane networks, for example, save lives for well under $100,000 each. But due to a mix of funding constraints, legislation, jurisdiction, and institutional inertia, DOT is not doing many of those projects.

Protected bike lane networks, for example, save lives for well under $100,000 each.

Link?

I wouldn't say that I've done a ton of research on this question but here are a couple of studies I've seen on this:

  • This report studied projects in Bogotá and Guangzhou and found that spending $1MM would save 2.27 or 1.38 lives per year (respectively). If you assume the infrastructure lasts 30 years that gives you $14,662 or $24,154 per life saved (respectively).
  • This study looked at bike lanes in NYC and basically found that the error bars are too wide to conclude anything; but their midpoint was $1297/QALY so perhaps somewhere around $50,000 per life saved.

The United States Department of Transportation will pay $11.8 million to save a life.

Flagging that this is the 2021 value. The most current, 2022, value is $12.5M.

great essay, thanks for writing + posting!

More from EJT
Curated and popular this week
 ·  · 25m read
 · 
Epistemic status: This post — the result of a loosely timeboxed ~2-day sprint[1] — is more like “research notes with rough takes” than “report with solid answers.” You should interpret the things we say as best guesses, and not give them much more weight than that. Summary There’s been some discussion of what “transformative AI may arrive soon” might mean for animal advocates. After a very shallow review, we’ve tentatively concluded that radical changes to the animal welfare (AW) field are not yet warranted. In particular: * Some ideas in this space seem fairly promising, but in the “maybe a researcher should look into this” stage, rather than “shovel-ready” * We’re skeptical of the case for most speculative “TAI<>AW” projects * We think the most common version of this argument underrates how radically weird post-“transformative”-AI worlds would be, and how much this harms our ability to predict the longer-run effects of interventions available to us today. Without specific reasons to believe that an intervention is especially robust,[2] we think it’s best to discount its expected value to ~zero. Here’s a brief overview of our (tentative!) actionable takes on this question[3]: ✅ Some things we recommend❌ Some things we don’t recommend * Dedicating some amount of (ongoing) attention to the possibility of “AW lock ins”[4]  * Pursuing other exploratory research on what transformative AI might mean for animals & how to help (we’re unconvinced by most existing proposals, but many of these ideas have received <1 month of research effort from everyone in the space combined — it would be unsurprising if even just a few months of effort turned up better ideas) * Investing in highly “flexible” capacity for advancing animal interests in AI-transformed worlds * Trying to use AI for near-term animal welfare work, and fundraising from donors who have invested in AI * Heavily discounting “normal” interventions that take 10+ years to help animals * “Rowing” on na
 ·  · 3m read
 · 
About the program Hi! We’re Chana and Aric, from the new 80,000 Hours video program. For over a decade, 80,000 Hours has been talking about the world’s most pressing problems in newsletters, articles and many extremely lengthy podcasts. But today’s world calls for video, so we’ve started a video program[1], and we’re so excited to tell you about it! 80,000 Hours is launching AI in Context, a new YouTube channel hosted by Aric Floyd. Together with associated Instagram and TikTok accounts, the channel will aim to inform, entertain, and energize with a mix of long and shortform videos about the risks of transformative AI, and what people can do about them. [Chana has also been experimenting with making shortform videos, which you can check out here; we’re still deciding on what form her content creation will take] We hope to bring our own personalities and perspectives on these issues, alongside humor, earnestness, and nuance. We want to help people make sense of the world we're in and think about what role they might play in the upcoming years of potentially rapid change. Our first long-form video For our first long-form video, we decided to explore AI Futures Project’s AI 2027 scenario (which has been widely discussed on the Forum). It combines quantitative forecasting and storytelling to depict a possible future that might include human extinction, or in a better outcome, “merely” an unprecedented concentration of power. Why? We wanted to start our new channel with a compelling story that viewers can sink their teeth into, and that a wide audience would have reason to watch, even if they don’t yet know who we are or trust our viewpoints yet. (We think a video about “Why AI might pose an existential risk”, for example, might depend more on pre-existing trust to succeed.) We also saw this as an opportunity to tell the world about the ideas and people that have for years been anticipating the progress and dangers of AI (that’s many of you!), and invite the br
 ·  · 3m read
 · 
Hi all, This is a one time cross-post from my substack. If you like it, you can subscribe to the substack at tobiasleenaert.substack.com. Thanks Gaslit by humanity After twenty-five years in the animal liberation movement, I’m still looking for ways to make people see. I’ve given countless talks, co-founded organizations, written numerous articles and cited hundreds of statistics to thousands of people. And yet, most days, I know none of this will do what I hope: open their eyes to the immensity of animal suffering. Sometimes I feel obsessed with finding the ultimate way to make people understand and care. This obsession is about stopping the horror, but it’s also about something else, something harder to put into words: sometimes the suffering feels so enormous that I start doubting my own perception - especially because others don’t seem to see it. It’s as if I am being gaslit by humanity, with its quiet, constant suggestion that I must be overreacting, because no one else seems alarmed. “I must be mad” Some quotes from the book The Lives of Animals, by South African writer and Nobel laureate J.M. Coetzee, may help illustrate this feeling. In his novella, Coetzee speaks through a female vegetarian protagonist named Elisabeth Costello. We see her wrestle with questions of suffering, guilt and responsibility. At one point, Elisabeth makes the following internal observation about her family’s consumption of animal products: “I seem to move around perfectly easily among people, to have perfectly normal relations with them. Is it possible, I ask myself, that all of them are participants in a crime of stupefying proportions? Am I fantasizing it all? I must be mad!” Elisabeth wonders: can something be a crime if billions are participating in it? She goes back and forth on this. On the one hand she can’t not see what she is seeing: “Yet every day I see the evidences. The very people I suspect produce the evidence, exhibit it, offer it to me. Corpses. Fragments of