Davidmanheim

Head of Research and Policy @ ALTER - Association for Long Term Existence and Resilience
7171 karmaJoined Working (6-15 years)

Participation
4

  • Received career coaching from 80,000 Hours
  • Attended more than three meetings with a local EA group
  • Completed the AGI Safety Fundamentals Virtual Program
  • Completed the In-Depth EA Virtual Program

Sequences
2

Deconfusion and Disentangling EA
Policy and International Relations Primer

Comments
850

Topic contributions
1

Close enough not to have any cyclic components that would lead to infinite cycles for the nonsatiable component of their utility.

Humans are neither coherent, nor do they necessarily have a nonsatiable goal - though some might. But they have both to a far greater extent than less intelligent creatures.

Are you willing to posit that advanced systems are coherent, with at least one non-satiable component? Because that's pretty minimal as an assumption, but implies with probability one that they prefer paperclipping of some sort.

Where and when are these supposed to occur, and how can we track that for our respective countries?

"EA has always bee[n] rather demanding,"

I want to clarify that this is a common but generally incorrect reading of EA's views. EA leaders have repeatedly clarified that you don't need to dedicate your life to it, and can simply donate to causes that others have identified as highly effective, and otherwise live your life.

If you want to do more than that, great, good for you - but EA isn't utilitarianism, so please don't confuse the demandingness of the two.

First, Utilitarianism doesn't traditionally require the type of extreme species neutrality you propose here.Singer and many EAs gave a somewhat narrower view of what 'really counts' as Utilitarian, but your argument assumes that narrow view without really justifying it.

Second, you assume future AIs will have rich inner lives that are valuable, instead of paperclipping the universe. You say "one would need to provide concrete evidence about what kinds of objectives advanced AIs are actually expected to develop" - but Eliezer has done that quite explicitly.

Very much in favor of posts clarifying that cause neutrality doesn't require value neutrality or deference to others' values.

I very much appreciate that you are thinking about this, and the writing is great. That said, without trying to address the arguments directly, I worry that the style here is justifying a conclusion you've come to and explores analogies you like rather than exploring the arguments and trying to decide what side to be on, and it fails to embrace scout mindset sufficiently to be helpful.

I think that replaceability is very high, so the counterfactual impact is minimal. But that said, there is very little possibility in my mind that even helping with RLHF for compliance with their "safety" guidelines is more beneficial for safety than for accelerating capabilities racing, so any impact is negative.

Load more