Suppose you believe AGI (or superintelligence) will be created in the future. In that case, you should also acknowledge its super capabilities in addressing EA problems like global health and development, pandemics, animal welfare, and cause prioritization decision-making.
Suppose you don't believe superintelligence is possible. In that case, you can continue pursuing other EA problems, but if you do believe superintelligence is coming, then why are you spending time and money on issues that will likely all be solved by AI, assuming superintelligence comes aligned with human values?
I've identified a few potential reasons why people continue to devote their time and money to non-AI-related EA causes:
- You aren't aware of the potential capabilities of superintelligence.
- You don't think that superintelligence will arrive for a long time, or you remain uncertain about a timeline.
- You're passionate about a particular cause, and superintelligence doesn't interest you.
- You believe that present suffering matters intrinsically, and that the suffering occurring now has a moral weight that can't be dismissed.
- You might even think that superintelligence won't be able to address particular problems.
It's widely believed (at least in the AI safety community) that the development of sufficiently advanced AI could lead to major catastrophes, a global totalitarian regime, or human extinction, all of which seem to me to be more pressing and critical than any of the above reasons for focusing on other EA issues. I post this because I'd like to see more time and money allocated to AI safety, particularly in solving the alignment problem through automated AI labor (since I don't believe human labor can solve it anytime soon, but that's beyond the scope of this post).
So, do any of the reasons presented above apply to you? Or do you have different reasons for not focusing on AI risks?
I don't claim you can align human groups with individual humans. If I'm reading you correctly, I think you're committing a category error in assigning alignment properties to groups of people like nation states or companies. Alignment, as I'm using the term, is the alignment of goals or values from an AI to a person or group of people. We expect this, I think, in part because we're accustomed to telling computers what to do and having them do exactly what we say (not always exactly what we mean, though).
Alignment is extremely tricky for the unenhanced human, but theoretically possible. My first best guess at solving it would be to automate the research and development of it with AI itself. We'll soon reach a sufficiently advanced AI that's capable of reasoning beyond anything anyone on Earth can come up with; we just have to ensure that the AI is aligned and that the one that trained that one is also aligned, and so on. My second-best guess would be through BCIs, and my third would be whole-brain emulation interpretability.
Assuming we even do develop alignment techniques, I'd argue that exclusive alignment (that is, for one or a small group of people) is more difficult than aligning with humanity.at large for the following reasons (I realize some of these go both ways, but I include them because I see them as more serious for exclusive alignment–like value drift):