Davidmanheim

Head of Research and Policy @ ALTER - Association for Long Term Existence and Resilience
7013 karmaJoined Working (6-15 years)

Participation
4

  • Received career coaching from 80,000 Hours
  • Attended more than three meetings with a local EA group
  • Completed the AGI Safety Fundamentals Virtual Program
  • Completed the In-Depth EA Virtual Program

Sequences
2

Deconfusion and Disentangling EA
Policy and International Relations Primer

Comments
813

Topic contributions
1

As an aside, the reports end up being critical in worlds where the rapid response is needed, since they show ongoing attention, and will be looked at in retrospect. But they can also be used more directly to galvanize news coverage on key topics and as evidence by policy orgs. Promoting that avenue for impact seems valuable.

Responding to only one minor point you made, the 6-month pause letter seems like the type of thing you oppose: it's not able to help with the risk, it just does deceptive PR that aligns with the goals of pushing against AI progress, while getting support from those who disagree with their actual goal.

I agree that there are drawbacks to interpersonal donations, but strongly disagree that building a complex overhead structure is a good idea. The problem exists, but the solution is worse.

I think it's useful to distinguish between industrial policy, regulation, and nationalization, and your new term seems to be somewhere in between. I think your model is generally useful, but at the same time, introducing a new term without being very clear about what it means in relation to existing terms is probably more confusing than clarifying.

If the claim was that this is best among public outreach interventions, the title is misleading. The post also doesn't really compare deep questioning to other public outreach methods, just justifies it on its own terms.

Opportunity costs for attention and time are the other things people could be doing, and it it common and I think basically justifiable to value people's time at a level similar to their work salary. The reasoning is that typically, even if you can't make money during your free time, people are willing to spend money and give up other opportunities to get free time - if they want to use that time to do deep questioning, that's great, but if and when they do, they are explicitly valuing that use of their time over other options.

And I agree that some grassroots organizations could push this forward, but I worry doing it on behalf of an organization with an explicit agenda, even as a volunteer, might undermine the personal connection of deep questioning. As you said, "the interlocutors do not have the impression that the public outreacher is from an organization and tries to persuade them of something." If they are, in fact, coming from an organization, that seems to be deeply deceptive.

I think this is a great thing to do, and have a general policy of approving of net positive actions. I also think that there are several convincing claims about why this is potentially tractable, including the arguments that it is fully additive, and that it avoids the need for coordination. Unfortunately, I also think that it fails to clear the bar for what we should expect in effective altruism in a couple ways.

First, I remain unconvinced that it's a "top-effective" intervention. It's unreasonable to say that it is cost-effective when there are opportunity costs which are not explored, and the actual impact is not quantified. As Vasco lays out below, there should be a stronger case that this is better than alternatives, or a clearer case that it should be done as an effective but ancillary activity which people can do in their free time, rather than as something that is as effective as the human league or other campaigns. To change that, I think that there should be a clear explanation of what is required for this to succeed in individual cases (training, experience,) an estimate of the time required, and an estimate of the actual impact (what proportion of people change their behavior, how much does it change, how long does the change persist,) and a exploration of how and when this could be net-negative (if done poorly, if it generates pushback when done at scale, etc.)

Second, I also think that it's not ambitious enough on its own terms; if it is as effective as claimed, how can it be scaled up effectively? Should there be volunteer training groups to teach people to do this more widely? Can this be done via existing networks? Could there be a trial designed to measure impact?

To conclude, overall, I think that this is admirable and a potentially tractable, but presented with misleadingly strong claims, and as I outlined above, neither as clear on several points as it could be, nor as ambitious as would be beneficial. 

This makes a number of non-trivial assumption and unsourced claims about a number of different issues, from relative moral value of animals to the carrying capacity of different biomes; I know that many of these are seen as common wisdom in EA, but I think failing to lay them out greatly weakens the conclusions.

Also, some questions to think about: Why are insects ignored? How does the transition happen, legally or economically? What are the impacts of land use changes, and do farmers sell the land? (To whom?) Do social norms around meat undermine the viability of a transition?

{making humanity more safe VS shortening AGI timelines} is itself a false dichotomy or false spectrum.

Why? Because in some situations, shortening AGI timelines could make humanity more safe, such as by avoiding an overhang of over-abundant computing resources that AGI could abruptly take advantage of if it’s invented too far in the future (the “compute overhang” argument).


I think this also ignores the counterfactual world with less safety research, where the equivalent advances, which are funded because of commercial incentives, come from less generalizable safety research, and we end up with less well prosaically aligned but similarly capable systems. (And I haven't really laid out this argument before, but I think it generalizes to the counterfactual world without OpenAI or even Deepmind being inspired by AI safety concerns.)

I think it's fine, if orgs are set up for getting the donations. If they have a "donate" button or page, they are set up to get the money, less credit card fees, etc. They problem is that setting up that sort of thing is anywhere from easy to legally very complex.

As someone who runs an organization that does a lot of biorisk work, it's incredibly expensive  in staff time and logistics to receive small donations - but if you're giving more than, say, $5,000, you could just email the organizations to ask, and I'm sure they could figure it out.

But as I answered, CHS does have a donation page. (And NTI does allow donations, with a box to indicate where you'd like the money to go, but it's unclear to me if that actually lets you direct it only to bio.)

Load more