Hide table of contents

I want to get a sense for what kinds of things EAs — who don't spend most of their time thinking about AI stuff — find most confusing/uncertain/weird/suspect/etc. about it.

By "AI stuff", I mean anything to do with how AI relates to EA.

For example, this includes:

  • What's the best argument for prioritising AI stuff?, and
  • How, if at all, should I factor AI stuff into my career plans?

but doesn't include:

  • How do neural networks work? (except inasmuch as it's relevant for your understanding of how AI relates to EA).

Example topics: AI alignment/safety, AI governance, AI as cause area, AI progress, the AI alignment/safety/governance communities, ...

I encourage you to have a low bar for writing an answer! Short, off-the-cuff thoughts very welcome.

25

0
0

Reactions

0
0
New Answer
New Comment

10 Answers sorted by

I've since gotten a bit more context, but I remember feeling super confused about these things when first wondering how much to focus on this stuff:

  1. Before we get to "what's the best argument for this," just what are the arguments for (and against) (strongly) prioritizing AI stuff (of the kind that people in the community are currently working on)?
    1. People keep saying heuristic-y things about self-improving AI and paperclips--just what arguments are they making? (What are the end-to-end / logically thorough / precise arguments here?)
    2. A bunch of people seem to argue for "AI stuff is important" but believe / act as if "AI stuff is overwhelmingly important"--what are arguments for the latter view?
    3. Even if AI is overwhelmingly important, why does this imply we should be focusing on the things the AI safety/governance fields are currently doing?
    4. Some of the arguments for prioritizing AI seem to route through "(emerging) technologies are very important"--what about other emerging technologies?
    5. If there's such a lack of strategic clarity / robustly good things to do in AI governance, why not focus on broadly improving institutions?
    6. Why should we expect advanced AI anytime soon?
  2. What are AI governance people up to? (I.e. what are they working on / what's their theory of change?)
  3. What has the AI safety field accomplished (in terms of research, not just field-building)? (Is there evidence that AI safety is tractable right now?)
  4. A lot of the additional things I found suspect were outside-view-y considerations / "common sense" heuristics--to put it in a very one-sided way, it was something like, "So you're telling me some internet forum is roughly the first and only community to identify the most important problem in history, despite this community's vibes of overconfidence and hero-worship and non-legible qualifications and getting nerd-sniped, and this supposedly critical problem just happens to be some flashy thing that lines up with their academic interests and sounds crazy and isn't a worry for the most clearly relevant experts?")

(If people are curious, the resources I found most helpful on these were: this, this, and this for 1.1, the former things + longtermism arguments + The Precipice on non-AI existential risks for 1.2, 1.1 stuff & stuff in this syllabus for 1.3 and 3, ch. 2 of Superintelligence for 1.4, this for 1.6, the earlier stuff (1.1 and 3) for 4, and various more scattered things for 1.5 and 2.)

(I have completely no expertise in AI, but this is what I always felt personally confused about)
How are we going to know/measure/judge whether our efforts to prevent AI risks are actually helping? Or how much they are helping? 

Firstly, thank you for this! For such a big priority (within the EA community), I feel like there's a lot of confusion about AI.

I help organize UChicago EA & have asked members to send me questions, so I'll update this comment as they come in:

  • AI Alignment: Do we need to decide on a moral principle(s) first? How would it be possible to develop beneficial AI without first 'solving' ethics/morality?
  • AI is neglected by the world generally but doesn’t seem neglected within EA. Does this have any implications for career planning?

Do we need to decide on a moral principle(s) first? How would it be possible to develop beneficial AI without first 'solving' ethics/morality?

Good question! The answer is no: 'solving' ethics/morality first is one thing that we probably eventually need to do, but we could first solve a narrower, simpler form of AI alignment, and use those aligned systems to help us solve ethics/morality and the other trickier problems (like the control problem for more general, capable systems). This is more or less what is discussed in ambitious vs narrow value learning. Narrow value learning is one narrower, simpler form of AI alignment. There are others, discussed here under the heading "Alternative solutions".

How worried are people actually about suffering in neural networks/artificial minds? 

(My impression is that this is a fun thing to talk about, but won't be that useful for a long time)

Here's a great post about this, which I would summarise as "not worried yet, but it's really hard to tell when we should worry".

Hi Sam. I'm curious to what extent people in the field think risk communication could be beneficial for reducing AI risk. In other words, are there any aspects of AI risk that could be mitigated by large numbers of people having accurate knowledge about them? Or is AI risk communication largely irrelevant to the problem? Or is it more likely to increase rather than decrease AI risk (perhaps by means of some type of infohazard)?

Here's a couple that came to mind just now.

  1. How smart do you need to be to contribute meaningfully t AI safety? Near top in class in high-school? Near top in class at ivy-league? Potential famous prof at ivy league? Potential fields medalist?

  2. Also, how hard should we expect alignment to be? Are we trying to throw resources at a problem we expect to be able to at least partially solve in most worlds (which is e.g. the superficial impression I get from biorisk) or are we attempting a hail mary, because it might just work and it's important enough to be worth a try (not saying that would be bad)?

  3. Big labs in the West that kind of target AGI are OpenAI and DeepMind. Others target AGI less explicitly, but inlcude e.g. Google Brain. Are there equivalents elsewhere? China? Do we know whether these exits? Am I missing labs that target AGI in the West?

  4. Finally, this one's kind of obvious, but how large is the risk? What's the probability of catastrophe? I'm aware of many estimates, but this is still definitely something I'm confused about.


I think on all these questions except (3), there's substantial disagreement among AI safety researchers, though I don't have a good feeling for the distributions of views either.

Thank you for posting this question and encouraging people to talk openly about this topic!

Here are some of the AI-related questions that I've thought about from time to time:

  • On the margin, should donors prioritize AI safety above other existential risks and broad longtermist interventions? Open Phil gives $10's of millions to AI safety and biosecurity every year, but other x-risks and longtermist areas seem rather unexplored and neglected, like s-risks.
  • What would make an artificial entity (like a computer program) sentient? What would count as a painful experience for said entity? Can we learn about this by studying the neuroscience of animal sentience?
  • In expectation, will there be more sentient artificial beings than sentient biological beings (including animals) over the long-term future? (brought up as an objection to this)
  • Is "intelligence" (commonly defined as the cognitive ability to make and execute plans to achieve goals) really enough to make an AI system more powerful than humans (individuals, groups, or all of humanity combined)?
  • Should we expect AI development to move toward AGI, narrowly superhuman AIs, CAIS, or something else?
  • What benefits and risks should we expect in a CAIS scenario?

On the margin, should donors prioritize AI safety above other existential risks and broad longtermist interventions?

To the extent that this question overlaps with Mauricio's question 1.2 (i.e. A bunch of people seem to argue for "AI stuff is important" but believe / act as if "AI stuff is overwhelmingly important"--what are arguments for the latter view?), then you might find his answer helpful.

other x-risks and longtermist areas seem rather unexplored and neglected, like s-risks

Only a partial answer, but worth noting that I think the most plausible... (read more)

Is "intelligence" ... really enough to make an AI system more powerful than humans (individuals, groups, or all of humanity combined)?

Some discussion of this question here: https://www.alignmentforum.org/posts/eGihD5jnD6LFzgDZA/agi-safety-from-first-principles-control

Here are some big and common questions I've received from early-stage AI Safety focused people, with at least some knowledge of EA.

They probably don't spend most of their time thinking about AIS, but it is their cause area of focus. Unsure if that meets the criteria you're looking for, exactly.

  1. What evidence would be needed for EA to depriotise AI Safety as a cause area, at least relative to other x-risks?
  2. What is the most impactful direction of research within AIS (common amongst people looking for their first project/opportunity. I usually point them at this lesswrong series, as a starting point).

I think that as a software developer I can't really help with this problem, but I'm not sure and would I'd like input from people in the field

A timely post: https://forum.effectivealtruism.org/posts/DDDyTvuZxoKStm92M/ai-safety-needs-great-engineers

(The focus is software engineering not development, but should still be informative.)

I'd like help vetting this role.

Here's my attempt

(Is this on-topic enough?)

[comment deleted]1
0
0
Curated and popular this week
 ·  · 25m read
 · 
Epistemic status: This post — the result of a loosely timeboxed ~2-day sprint[1] — is more like “research notes with rough takes” than “report with solid answers.” You should interpret the things we say as best guesses, and not give them much more weight than that. Summary There’s been some discussion of what “transformative AI may arrive soon” might mean for animal advocates. After a very shallow review, we’ve tentatively concluded that radical changes to the animal welfare (AW) field are not yet warranted. In particular: * Some ideas in this space seem fairly promising, but in the “maybe a researcher should look into this” stage, rather than “shovel-ready” * We’re skeptical of the case for most speculative “TAI<>AW” projects * We think the most common version of this argument underrates how radically weird post-“transformative”-AI worlds would be, and how much this harms our ability to predict the longer-run effects of interventions available to us today. Without specific reasons to believe that an intervention is especially robust,[2] we think it’s best to discount its expected value to ~zero. Here’s a brief overview of our (tentative!) actionable takes on this question[3]: ✅ Some things we recommend❌ Some things we don’t recommend * Dedicating some amount of (ongoing) attention to the possibility of “AW lock ins”[4]  * Pursuing other exploratory research on what transformative AI might mean for animals & how to help (we’re unconvinced by most existing proposals, but many of these ideas have received <1 month of research effort from everyone in the space combined — it would be unsurprising if even just a few months of effort turned up better ideas) * Investing in highly “flexible” capacity for advancing animal interests in AI-transformed worlds * Trying to use AI for near-term animal welfare work, and fundraising from donors who have invested in AI * Heavily discounting “normal” interventions that take 10+ years to help animals * “Rowing” on na
 ·  · 3m read
 · 
About the program Hi! We’re Chana and Aric, from the new 80,000 Hours video program. For over a decade, 80,000 Hours has been talking about the world’s most pressing problems in newsletters, articles and many extremely lengthy podcasts. But today’s world calls for video, so we’ve started a video program[1], and we’re so excited to tell you about it! 80,000 Hours is launching AI in Context, a new YouTube channel hosted by Aric Floyd. Together with associated Instagram and TikTok accounts, the channel will aim to inform, entertain, and energize with a mix of long and shortform videos about the risks of transformative AI, and what people can do about them. [Chana has also been experimenting with making shortform videos, which you can check out here; we’re still deciding on what form her content creation will take] We hope to bring our own personalities and perspectives on these issues, alongside humor, earnestness, and nuance. We want to help people make sense of the world we're in and think about what role they might play in the upcoming years of potentially rapid change. Our first long-form video For our first long-form video, we decided to explore AI Futures Project’s AI 2027 scenario (which has been widely discussed on the Forum). It combines quantitative forecasting and storytelling to depict a possible future that might include human extinction, or in a better outcome, “merely” an unprecedented concentration of power. Why? We wanted to start our new channel with a compelling story that viewers can sink their teeth into, and that a wide audience would have reason to watch, even if they don’t yet know who we are or trust our viewpoints yet. (We think a video about “Why AI might pose an existential risk”, for example, might depend more on pre-existing trust to succeed.) We also saw this as an opportunity to tell the world about the ideas and people that have for years been anticipating the progress and dangers of AI (that’s many of you!), and invite the br
 ·  · 3m read
 · 
Hi all, This is a one time cross-post from my substack. If you like it, you can subscribe to the substack at tobiasleenaert.substack.com. Thanks Gaslit by humanity After twenty-five years in the animal liberation movement, I’m still looking for ways to make people see. I’ve given countless talks, co-founded organizations, written numerous articles and cited hundreds of statistics to thousands of people. And yet, most days, I know none of this will do what I hope: open their eyes to the immensity of animal suffering. Sometimes I feel obsessed with finding the ultimate way to make people understand and care. This obsession is about stopping the horror, but it’s also about something else, something harder to put into words: sometimes the suffering feels so enormous that I start doubting my own perception - especially because others don’t seem to see it. It’s as if I am being gaslit by humanity, with its quiet, constant suggestion that I must be overreacting, because no one else seems alarmed. “I must be mad” Some quotes from the book The Lives of Animals, by South African writer and Nobel laureate J.M. Coetzee, may help illustrate this feeling. In his novella, Coetzee speaks through a female vegetarian protagonist named Elisabeth Costello. We see her wrestle with questions of suffering, guilt and responsibility. At one point, Elisabeth makes the following internal observation about her family’s consumption of animal products: “I seem to move around perfectly easily among people, to have perfectly normal relations with them. Is it possible, I ask myself, that all of them are participants in a crime of stupefying proportions? Am I fantasizing it all? I must be mad!” Elisabeth wonders: can something be a crime if billions are participating in it? She goes back and forth on this. On the one hand she can’t not see what she is seeing: “Yet every day I see the evidences. The very people I suspect produce the evidence, exhibit it, offer it to me. Corpses. Fragments of