Hide table of contents

Other titles I considered:

  • Lord Martin Rees’ new existential risk book published: "If Science is to Save Us" 
  • Lord Martin Rees should be on more existential risk introductions
  • Lord Martin Rees is a boss

[Disclaimer: I work at CSER, which Martin co-founded (in some sense he’s my boss), and I know and like him, so I’m biased towards him. However, I’ve mostly referred to objective, verifiable evidence, and others can corroborate.]

Summary: 

Lord Martin Rees is one of the UK’s, and perhaps the world’s, most eminent scientists, has been an advocate of existential risk since at least 2003 (and arguably since the 1970s), and is a charismatic speaker and engaging writer. I’ll argue he should be featured in more ‘introductions to existential risk’ and should be turned to as a powerful advocate for existential risk reduction - like e.g. Will MacAskill and Toby Ord. In this short piece I’ll give a quick bio and describe some of his work. In part, this post is just an appreciation post – he’s just recently had his 80th birthday, after all. 


Photo by Hanna-Katrina Jędrosz for the New Statesman

New book

The occasion for this post is that his new book has just been published: If Science is to Save Us. Summary:

There has never been a time when ‘following the science’ has been more important for humanity. At no other point in history have we had such advanced knowledge and technology at our fingertips, nor had such astonishing capacity to determine the future of our planet.

But the decisions we must make on how science is applied belong outside the lab and should be the outcome of wide public debate. For that to happen, science needs to become part of our common culture. Science is not just for scientists: if it were, it could never save us from the multiple crises we face. For science can save us, if its innovations mesh carefully into society and its applications are channelled for the common good.

As Martin Rees argues in this expert and personal analysis of the scientific endeavour on which we all depend, we need to think globally, we need to think rationally and we need to think long-term, empowered by twenty-first-century technology but guided by values that science alone cannot provide.

Coverage:

 

Quick bio

Martin Rees is a cosmologist and astrophysicist who’s done leading research (500+ papers) on black holes, quasars and the multiverse. He knew Stephen Hawking well (and wrote a very nice obituary for him).

He’s been the Astronomer Royal (previous holders, Halley of ‘Halley’s Comet’ fame) since 1995. He was the 60th President of the Royal Society, 2005-2010 (previous holders Wren, Pepys, Newton, Rutherford, etc). The Royal Society is the UK’s national academy of sciences, and one of the most preeminent in the world. He was Master of Trinity College, Cambridge 2004-2012. He was made a Lord, specifically a non-party-political (‘crossbench’) member of the House of Lords, in 2005. He’s published 10+ books including:

  • From Here to Infinity: Scientific Horizons (UK) / From Here to Infinity: A Vision for the Future of Science 
  • Just Six Numbers
  • Our Cosmic Habitat
  • Before the Beginning: Our Universe and Others
  • Cosmic Coincidences: Dark matter, mankind and anthropic cosmology
  • Gravity’s Fatal Attraction: Black Holes in the Universe
  • New Perspectives in Astrophysical Cosmology

This is all to say he’s one of the UK’s (and perhaps the world) leading scientists. 

 

Good introductory materials

Rees has given two TED talks, which have together been watched 4-5 million times. I think they’re great introductions to the subject.

Can we prevent the end of the world? 

1.4 million views across YouTube and Ted website.

   

Earth in its final century? 

3.9 million views across YouTube and Ted website.

 

Several of his books are specifically on existential risk:

I think these should be included on ‘introduction to existential risk’ reading lists as very engaging and credible intros.

 

Notable contributions to existential risk research

Martin Rees has been a major populariser of existential risk - and has also made core intellectual contributions.

 

General

Most importantly, Martin was really early to existential risk - ‘Our Final Century’ was published in 2003, around the time Nick Bostrom was getting going.

He co-founded the Centre for the Study of Existential Risk at Cambridge University in 2012 - and helped advise on the founding of other centres, e.g. the Future of Life Institute in 2014. He was an early sponsor of the APPG for Future Generations. He played an important role in setting up the Lords Select Committee on Risk Assessment and Risk Planning, which published the 2021 report Preparing for Extreme Risks: Building a Resilient Society.

As Astronomer Royal, his views on space carry particular weight. He’s the most prominent critic of the argument that colonising Mars is a good way to reduce existential risk (see e.g. GuardianVox). He’s also closely involved with SETI, where he’s been one of the leading voices suggesting that any signs of extraterrestrial life are far more likely to be digital/artificial than biological, and this should shape SETI’s search.

Long before 2003, he was involved in the Pugwash Conferences in the 1970s - one of the only opportunities for scientists from either side of the Iron Curtain to engage with one another. Martin sometimes describes the current generation of existential risk activists and researchers as perhaps the third wave of concern around existential risk. First the ‘concerned scientists’ that participated in the Manhattan Project, set up the Bulletin of the Atomic Scientists and successfully ‘tabooed’ the use of nuclear weapons in the 1950s. This is the generation he looked up to. Then the second, the Pugwash scientists of the 1960s and 70s that contributed to the arms control agreements: the Biological Weapons Convention and the Strategic Arms Limitation Talks (SALT) treaties. I find this sense of history and tradition very motivating. We didn’t come up with all of this all on our own: we’re part of a wider story and can build on their successes.


Climate change

One particularly notable example on climate is his work at the Vatican. In May 2014, he helped Sir Partha Dasgupta co-organise a major workshop with the Pontifical Academy of Sciences on climate change. After the workshop, Sir Partha spoke to the Pope directly and encouraged him to include climate change in his speeches and to urge people to be better stewards of the planet. The workshop underpinned a major report published in April 2015 by the Vatican. The report in turn partly informed the May 2015 Laudato si’ Papal Encyclical, which focussed on the impending threat of climate change and was influential in encouraging the 1.3 billion Catholics worldwide to support for the Paris Agreement, agreed in December 2015.

Martin also contributed to A Global Apollo Programme To Combat Climate Change (June 2015), an early and prominent call for major R&D into clean energy, now widely seen as perhaps the most important contribution philanthropists and Western governments can make; and Climate Change: A Risk Assessment (July 2015), an early and prominent analysis of worst case scenarios which again encouraged the Paris Agreement.

None of this is to say he was one of the most important figures, simply to say he was an early and prominent advocate.
 

Biorisk

Martin was also an early and prominent voice warning about biorisks. For example, in 2017 he made a famous bet (described as the bet of the century) with Stephen Pinker that “A bioterror or bioerror will lead to one million casualties in a single event within a six month period starting no later than Dec 31 02020.” (It has not resolved yet, due to lack of clarity on whether Covid-19 was a lab leak or natural emergence.)

 

Personal

On a personal (almost gushy) note, Martin is kind, supportive, and encouraging to junior colleagues. Lots of CSER staff, and former Cambridge students, will have stories of his warmth and support.

Many senior academics or thought-leaders can be spiky and dismissive to junior colleagues. This is true of academia general, and unfortunately can sometimes be true in existential risk and AI alignment. I think Martin shows this is completely unnecessary. He’s got about as high up as one can get in science and academia - and is just really nice. Martin is a great role model for the field.

Comments19


Sorted by Click to highlight new comments since:

On the face of it, Rees has indeed been neglected in introductory materials on this topic. Any idea why?

When I read 'On the Future' I wasn't that much of a fan. My memory is pretty hazy of it because I read it a few years ago, but my feeling about it was that it wasn't very content dense and some of it didn't seem quite correct to me. I was reading expecting to want to recommend it to others, but actually decidedly preferred the Precipice (and they cover fairly similar things, so I'm unlikely to recommend both). Not sure if others feel that way, but if they do it might contribute to (at least this book) being relatively less mentioned in introductory materials.

Tbc, I'm still a big fan of Lord Martin Rees' work!

I agree that if I could only recommend one book, it would probably be the Precipice  - as its more up-to-date and comprehensive. I was thinking a wider bibliography / reading list. However, I really would prioritise the two Ted talks as short, interesting, credible intros.

I've got a real soft spot for "Our Final Century: Will Civilisation Survive the Twenty-first Century?" as it was the book that originally got me interested in existential risk. I still think its really important for the field, and is usefully included alongside 2008's GCR and Bostrom's 2002 paper. We're actually working on an "updated after 20 years" version of the book, hopefully out next year.

Probably just Oxford vs Cambridge founder effects/path dependency. 

EDIT: By 'Oxford founder effects' I mean something like "many of the early xrisk researchers came up through Oxford & naturally tend to cite Oxford folks; and two of the best book-length recent intros are from Toby and Will at Oxford; so the introductory materials are skewed towards Oxford".

[anonymous]17
4
7

He  also provided a blurb for Emile Torres' book, well after Torres said that Nick Bostrom, Will MacAskill, Eliezer Yudkowsky, Toby Ord, Hilary Greaves etc endorse white supremacist ideology and eugenics. 

Can you give a link for verifying this claim?

Also, I'd be pretty surprised if this were the reason for EA avoiding heavy promotion of Rees's work.

https://www.xriskology.com/books

I guess if that were the reason it'd probably be because people worry that implies Rees might agree with a bunch of Torres' views they think are very bad. though I think that forwarding someone's book or blurbing someone's book is pretty consistent with disagreeing strongly with a bunch of their stuff (if you even know about it).

Rees has also written multiple blurbs for Will MacAskill, Nick Bostrom et al.

unsure why downvoted. upvoted for being a possible reason he might not be mentioned more (not saying it's a good reason ).

sidenote: if we're so parochial that Cambridge is too far for Oxford-doninated ea to take notice of what goes on there.... that seems like pretty bad news.

[anonymous]12
7
0

Oxford vs Cambridge seems more likely to me than the blurb explanation because Torres' book was published in 2017 and would only explain changes after that time, but I don't have any particular reason to think anything changed at that time. Happy to be corrected though. 

FWIW I read his book On the Future in 2018 and thought it was mediocre. The abundance of 2- and 3-star reviews on Goodreads reflect my recollection of it.

Sharing this not to be critical of him (I'm not going to judge him or his work based on this one book), but to recommend anyone here who is new to his work start with something other than On the Future.

In a conversation with David Deutsch at the RSA, Rees said something which surprised me:

My experience is that advice from experts to politicians directly is not heeded. It is far better if the experts get through to the public and the press and then there is pressure from MPs postbags and the press, and politicians do respond to that. So that is another reason why I think it is important that scientific experts should engage with the wider public because that is a way of having more influence.

My guess is that he’s wrong about this. But he’s spent quite a while being a Lord, and I have never spoken to an MP, so I feel I have to increase my credence on this at least a bit…

Yes he's said this very consistently for years. Its interesting for like the insider's insider (Astronomer Royal!) to advocate an outside game. 

As for everything, I suspect the answer is "you need both, determining which is most helpful on the margin depends on the specific details of each case and your own personal fit".

Beat me to it,  pressure from bottom up and top down are probably both needed. Also, could we say in a way that most politicians (those that don't directly influence the  given topic) can also be considered technically part of the wider public since they are media consumers.  Sometimes this can lead to sideways pressure.

His 'Just Six Numbers' is great too, on the apparent fine-tuning of cosmological constants for life.

The Sky at Night did a fantastic one-hour interview with him for their  June show, to celebrate his 80th birthday (unfortunately I'm not sure if the interview is accessible to those outside the UK). 

The focus is obviously astronomy, so mostly covers his achievements and interesting things that have happened in his lifetime. However, the final 20 minutes or so discusses extraterrestrial intelligence, the future of life/humanity, and associated threats.

The EA Handbook has a section called Our Final Century? which I assume is a homage to Martin Rees' book of the same title, though I don't believe there's any explicit mention of Rees' work.

Thanks for sharing. I've only been vaguely familiar with Rees' work and have been somewhat worried about largely deferring to (and recommending) a narrow and very homogenous set of people. I'm reminded of this syllabus on longtermism which includes a more diverse set of thinkers.

Curated and popular this week
 ·  · 25m read
 · 
Epistemic status: This post — the result of a loosely timeboxed ~2-day sprint[1] — is more like “research notes with rough takes” than “report with solid answers.” You should interpret the things we say as best guesses, and not give them much more weight than that. Summary There’s been some discussion of what “transformative AI may arrive soon” might mean for animal advocates. After a very shallow review, we’ve tentatively concluded that radical changes to the animal welfare (AW) field are not yet warranted. In particular: * Some ideas in this space seem fairly promising, but in the “maybe a researcher should look into this” stage, rather than “shovel-ready” * We’re skeptical of the case for most speculative “TAI<>AW” projects * We think the most common version of this argument underrates how radically weird post-“transformative”-AI worlds would be, and how much this harms our ability to predict the longer-run effects of interventions available to us today. Without specific reasons to believe that an intervention is especially robust,[2] we think it’s best to discount its expected value to ~zero. Here’s a brief overview of our (tentative!) actionable takes on this question[3]: ✅ Some things we recommend❌ Some things we don’t recommend * Dedicating some amount of (ongoing) attention to the possibility of “AW lock ins”[4]  * Pursuing other exploratory research on what transformative AI might mean for animals & how to help (we’re unconvinced by most existing proposals, but many of these ideas have received <1 month of research effort from everyone in the space combined — it would be unsurprising if even just a few months of effort turned up better ideas) * Investing in highly “flexible” capacity for advancing animal interests in AI-transformed worlds * Trying to use AI for near-term animal welfare work, and fundraising from donors who have invested in AI * Heavily discounting “normal” interventions that take 10+ years to help animals * “Rowing” on na
 ·  · 3m read
 · 
About the program Hi! We’re Chana and Aric, from the new 80,000 Hours video program. For over a decade, 80,000 Hours has been talking about the world’s most pressing problems in newsletters, articles and many extremely lengthy podcasts. But today’s world calls for video, so we’ve started a video program[1], and we’re so excited to tell you about it! 80,000 Hours is launching AI in Context, a new YouTube channel hosted by Aric Floyd. Together with associated Instagram and TikTok accounts, the channel will aim to inform, entertain, and energize with a mix of long and shortform videos about the risks of transformative AI, and what people can do about them. [Chana has also been experimenting with making shortform videos, which you can check out here; we’re still deciding on what form her content creation will take] We hope to bring our own personalities and perspectives on these issues, alongside humor, earnestness, and nuance. We want to help people make sense of the world we're in and think about what role they might play in the upcoming years of potentially rapid change. Our first long-form video For our first long-form video, we decided to explore AI Futures Project’s AI 2027 scenario (which has been widely discussed on the Forum). It combines quantitative forecasting and storytelling to depict a possible future that might include human extinction, or in a better outcome, “merely” an unprecedented concentration of power. Why? We wanted to start our new channel with a compelling story that viewers can sink their teeth into, and that a wide audience would have reason to watch, even if they don’t yet know who we are or trust our viewpoints yet. (We think a video about “Why AI might pose an existential risk”, for example, might depend more on pre-existing trust to succeed.) We also saw this as an opportunity to tell the world about the ideas and people that have for years been anticipating the progress and dangers of AI (that’s many of you!), and invite the br
 ·  · 12m read
 · 
I donated my left kidney to a stranger on April 9, 2024, inspired by my dear friend @Quinn Dougherty (who was inspired by @Scott Alexander, who was inspired by @Dylan Matthews). By the time I woke up after surgery, it was on its way to San Francisco. When my recipient woke up later that same day, they felt better than when they went under. I'm going to talk about one complication and one consequence of my donation, but I want to be clear from the get: I would do it again in a heartbeat. Correction: Quinn actually donated in April 2023, before Scott’s donation. He wasn’t aware that Scott was planning to donate at the time. The original seed came from Dylan's Vox article, then conversations in the EA Corner Discord, and it's Josh Morrison who gets credit for ultimately helping him decide to donate. Thanks Quinn! I met Quinn at an EA picnic in Brooklyn and he was wearing a shirt that I remembered as saying "I donated my kidney to a stranger and I didn't even get this t-shirt." It actually said "and all I got was this t-shirt," which isn't as funny. I went home and immediately submitted a form on the National Kidney Registry website. The worst that could happen is I'd get some blood tests and find out I have elevated risk of kidney disease, for free.[1] I got through the blood tests and started actually thinking about whether to do this. I read a lot of arguments, against as well as for. The biggest risk factor for me seemed like the heightened risk of pre-eclampsia[2], but since I live in a developed country, this is not a huge deal. I am planning to have children. We'll just keep an eye on my blood pressure and medicate if necessary. The arguments against kidney donation seemed to center around this idea of preserving the sanctity or integrity of the human body: If you're going to pierce the sacred periderm of the skin, you should only do it to fix something in you. (That's a pretty good heuristic most of the time, but we make exceptions to give blood and get pier