"I feel like if you're going to put something out there in the public sphere as a leader in AI, a bit of timeline conservatism might be prudent."
I see and respect that position, but you can imagine someone saying the opposite: "I feel like if you're going to put something out there in the public sphere as a leader in AI, it's probably prudent to warn people of significant risks that happens much sooner than people expect, even if you think it's less than 50% likely to happen then."
Plausibly you can get away with reporting 3-5 numbers.
For 3 numbers, 25th percentile, median, 75th percentile. This is the approach ("interquartile range") used for reporting SAT acceptance ranges in the US. So we have at least a prior example of a widely reported figure that people don't think "normal people"/high-schoolers and their parents would have too much trouble understanding.
For 5 numbers, something like 5th percentile, 25th percentile, median, 75th percentile, 95th percentile.
3-5 numbers obviously harder to communicate than 1 number, and less precise than the full distribution. But hopefully it's clear and useful enough to be good here.
I think they were optimizing for a combination of concreteness (so there's an exact story to point to, where the 2027 story is "things go roughly as they expected" whereas 2028 and 2031 were pricing in different types of individually unexpected delays[1]), and for memetic value.
Compare: My best estimate is that this project will take me 6 months. However, if you ask me to write it out step-by-step, it'd take me 4 months.The 6 months include buffers for various delays, some expected and some unexpected.
I think for project time estimations as part of a larger plan, the 6 month reply is more useful. But for someone following along on my thinking process, or a manager/collegue/direct report trying to help me optimize, the 4 month step-by-step report might be easier to follow along and/or more useful to critique or improve.
Agreed. It's possible that we/our descendants won't see much value for extending past blissful experiences even when other axes of value are theoretically possible, in the same way that aliens without conscious experiences would not see any particular reason to privilege qualia (even if they could be convinced that it's real).
I think a common mistake for researchers/analysts outside of academia[1] is that they don't focus enough on trying to make their research popular. Eg they don't do enough to actively promote their research, or writing their research in a way that's easy to become popular. I talked to someone (a fairly senior researcher) about this, and he said he doesn't care about mass outreach given that only cares about his research being built upon by ~5 people. I asked him if he knows who those 5 people are and could email them; he said no.
I think this is a systematic mistake most of the time. It's true that your impact often routes through a small number of people. However, only some of the time would you know who the decisionmakers are ahead of time (eg X philanthropic fund should fund Y project, B regulator should loosen regulations in C domain), and have a plan for directly reaching them. For the other cases, you probably need to reach at minimum thousands of vaguely-related/vaguely-interested people before the ~5 most relevant people for your research would come across your research.
Furthermore, popularity has other advantages:
Now of course it's possible to aim too much for popularity, and Goodhart on that. For example, by focusing on research topics that's popular rather than important, or on research directions/framings that's memetically fit rather than correct. Obsessing over metrics can also be bad for having the space to explore newer and more confusing ideas.
Nonetheless, on balance I think most researchers should be aware of what makes their research popular and in part gravitate towards that. Maybe I think they should spend >10% of their time on publicizing their work (not including "proof of work" style paper writing, grant applications, etc), whereas instead many ppl seem to spend <5%.
[1] Academia (and for that matter, for-profit research within a company) has this problem less because usually your peer group and potential collaborators in your sub-subfield are more well-defined and known to you. Also, academics care less about impact. Even so, I think ppl are leaving impact (and possibly career success) on the table by not being more popular. Eg if you work in theoretical econ you should aspire to have your theories be applied by applied economists, if you work on the evolutionary dynamics of bees you should want to be read by people working on ants, if you work on themes in Renaissance art history you should aspire to be read by people studying Renaissance political philosophy, etc.
This means (imo) academics should be more willing to have academic blogs and Twitter threads, and tolerate (or even seek out) media coverage of their work.
"There are more things in heaven and earth, Horatio, Than are dreamt of in your philosophy"
One thing I've been floating about for a while, and haven't really seen anybody else deeply explore[1], is what I call "further moral goods": further axes of moral value as yet inaccessible to us, that is qualitatively not just quantitatively different from anything we've observed to date.
For background, I think normal, secular, humans live in 3 conceptually distinct but overlapping worlds:
For the purposes of this post, I'm not that interested in the delineating between whether these worlds are truly different or just conceptually interesting ways to talk about things (ie I'm not positing a strong position on mathematical platonism or consciousness dualism)
But what's interesting to me is how these different worlds ground morality/value, what some philosophers would call "axiology." When people try to solely ground morality in the first two worlds, and even more so when people try to ground morality in the first world alone[3], deep believers in all three worlds (which I think is most people, and most philosophers) think they're entirely missing the point! It seems almost self-evident that conscious experience is much more important than the arrangement of mere rocks, or bloodless abstract game theory of feeling-less zombies!
But are these the only 3 worlds? Is it possible to have other morally relevant worlds, and in particular worlds that will self-evidently seem so much more important than subjective experience if only we know about them?
Perhaps.
For example, (most) religious people believe they have an answer, :
Now I think the religious people are wrong about the world as we see it today. But do we have strong reason to think that the three worlds as we know them are the only ones left? I think no.
In particular, we have two distinct reasons to think future intelligences can discover other worlds:
A. AIs, including future AIs, will be a distinct type of mind(s) than human mind(s). Just as most people today believe that humans (and other animals) have qualia that present-day AIs do not have, we should also think it's plausible that different mental architectures in AI will allow them to have moral goods that we cannot experience or perhaps even conceive.
B. Superintelligences (likely digital intelligences, though in theory could also be our posthuman descendants) will be able to search for further moral goods. At some point in the future (if we don't all die first), it will become trivial to spend more brainpower than has ever existed in all of human science and philosophy combined to search for other sources of moral value. This can come from engineering unique environmental arrangements of matter, unique structures of minds, or something else entirely.
So one day our descendants may discover worlds five, six, and so on: sources of moral value qualitatively distinct and superior to what we have access to, in the same way that grounding morality purely in game theory or entropy feels foolish to most experiencing humans today.
If true, this is a big deal! [4]
This seems overall quite possible to me. But is it probable?
I don't have a good sense of high likely this all is. Trying to estimate it feels beyond my forecasting or philosophical competence. But it seems plausible enough, and interesting enough, that I wanted to bring it to people's attention, in case other people have ideas on how to extend it.
Appendix A:
Existing literature: This concept is widespread but undertheorized. Mill's qualitative distinction among pleasures can point us in this direction; Bostrom's "Letter from Utopia" is the most vivid articulation ("What I feel is as far beyond feelings as what I think is beyond thoughts"); Danaher (2021) coined "axiological possibility space"; Ord's The Precipice argues we have "barely begun the ascent" and our investigations of flourishing may be "like astronomy before telescopes." According to a search from Claude, Nagel, Jackson, and Chalmers "collectively demonstrate that the space of possible conscious experiences vastly exceeds human experience." Banks's concept of Subliming: where "the very ideas, the actual concepts of good, of fairness and of justice just ceased to matter", is the most philosophically precise depiction I've seen in science fiction.
[1] Though I've seen shades of it in academic philosophy, EA/longtermist writing, science fiction/fantasy, and discussions of religion
[2] This is disputed.
[3] eg entropy as the guiding factor of morality, a la Beff Jezos.
[4] And if false, but convincing enough to be an attractor state for our descendants, this will sadly also be a very big deal.
On the off chance anybody is both interested in AI news and missed it, Anthropic sued DoW and other government officials/agencies for the supply chain risk designation in DC and Northern Californian Circuit. The full-text of the Northern Californian complaint here:
The primary complaints:
IANAL etc. in my personal opinion #2 seems very clearcut as a common-language and precedent reading of these things. #1 also seems strong. Sources I randomly skimmed online thought #3-#5 had a good case too, but I don't have an independent view.
The DC complaint looks less meaty (and I didn't read it)
I think they were laughed at enough after the Wired article (from here and elsewhere) that maintaining the previous line was no longer tenable for them.
I also separately think their current stated position is more accurate than the previous one, but I'm just observing that the incentives are a larger fraction of the story behind what they're saying than what ppl might otherwise be reading them as.
Just the 25th and 75th percentile?