DM

David_Moss

Principal Research Director @ Rethink Priorities
8665 karmaJoined Working (6-15 years)

Bio

I am the Principal Research Director at Rethink Priorities. I lead our Surveys and Data Analysis department and our Worldview Investigation Team. 

The Worldview Investigation Team previously completed the Moral Weight Project and CURVE Sequence / Cross-Cause Model. We're currently working on tools to help EAs decide how they should allocate resources within portfolios of different causes, and to how to use a moral parliament approach to allocate resources given metanormative uncertainty.

The Surveys and Data Analysis Team primarily works on private commissions for core EA movement and longtermist orgs, where we provide:

  • Private polling to assess public attitudes
  • Message testing / framing experiments, testing online ads
  • Expert surveys
  • Private data analyses and survey / analysis consultation
  • Impact assessments of orgs/programs

Formerly, I also managed our Wild Animal Welfare department and I've previously worked for Charity Science, and been a trustee at Charity Entrepreneurship and EA London.

My academic interests are in moral psychology and methodology at the intersection of psychology and philosophy.

How I can help others

Survey methodology and data analysis.

Sequences
3

RP US Public AI Attitudes Surveys
EA Survey 2022
EA Survey 2020

Comments
582

Yeh, I definitely agree that asking multiple questions per object of interest to assess reliability would be good. But also agree that this would lengthen a survey that people already thought was too long (which would likely reduce response quality in itself). So I think this would only be possible if people wanted us to prioritise gathering more data about a smaller number of questions.

Fwiw, for the value of hires questions, we have at least seen these questions posed in multiple different ways over the years (e.g. here) and continually produce very high valuations. My guess is that, if those high valuations are misleading, this is driven more by factors like social desirability than difficulty/conceptual confusion. There are some other questions which have been asked in different ways across years (we made a few changes to the wording this year to improve clarity, but aimed to keep the same where possible), but I've not formally assessed how those results differ. 

Thanks Vasco!

This bullet plus the other I quoted above suggest typical junior and senior hires have lifetimes of 40.2 (= 2.04*10^6/(50.7*10^3)) and 16.1 roles (= 7.31*10^6/(455*10^3)), which are unreasonably long. For 3 working-years per junior hire, and 10 working-years per senior hire, they would correspond to working at junior level for 121 years (= 40.2*3), and at senior level for 161 years (= 16.1*10).

We took a different approach to this here, where we looked at the ratio between the value people assigned to a role being filled at all and the value of a person joining the community, rather than the value of the first vs second most preferred hire.

If we look at those numbers, we only get a ratio of ~5 (for both junior and senior hires), i.e. however valuable people think a role being filled is, they think the value of getting a 'hire-level' person to the community is approximately 5x this. 

This seems more in line with the number of additional roles that we might imagine a typical hire goes onto after being hired for their first role. That said, people might also have been imagining (i) that people's value produced increases (perhaps dramatically) after their first role, (ii) that people create value for the community outside the roles they're hired to. 

Thanks for the comment Jessica! This makes sense. I have a few thoughts about this:

  • More time for people to answer, and in particular to reflect, sounds like it could have been useful (though I wasn't at the event, so I can't comment on the tradeoffs here).
  • My impression is that the difficulty of the survey is mostly due to the inherent difficulty of the questions we were asked to elicit judgements about (either/both because the questions were substantively difficult and required a lot of information/reflection- e.g. what is the optimal growth rate for EA- or because they're very conceptually difficult/counterintuitive- e.g. how much value do you assign to x relative to y controlling for the value of x's converting into y's), and less because of the operationalization of the questions themselves (see the confusion about earlier iterations of the questions).
    • One possible response to this, which was mentioned in feedback, is that it could be valuable to pose these questions to dedicated working groups, who devote extensive amounts of time to deliberating on them. Fwiw, this sounds like a very useful (though very costly) initiative to me. It would also have the downside of limiting input to an even smaller subset of the community: so perhaps ideally one would want to pose these questions to a dedicated group, presenting their findings to the wider MCF audiences, and then ask the MCF audience for their take after hearing the working group's findings. Of course, this would take much more time from everyone, so it wouldn't be valuable.
    • Another possible response is to just try to elicit much simpler judgements. For example, rather than trying to actually get a quantitative estimate of "how many resources do you think think each cause should get?", we could just ask "Do you think x should get more/less?" I think the devil is in the details here, and it would work better for some questions than others e.g. in some cases, merely knowing whether people think a cause should get more/less would not be action-guiding for decisionmakers, but in other cases it would (we're entirely dependent on what decisionmakers tell us they want to elicit here, since I see our role as designing questions to elicit the judgements we're asked for, not deciding what judgements we should try to elicit). 

Hey Manuel,

I think the public posts should start coming out pretty soon (within the next couple of weeks). 

That said I would strongly encourage movement builders and other decision-makers to reach out to us directly and request particular results when they are relevant to your work. We can often produce and share custom analyses within a day (much faster than a polished public post).

Many people believe that AI will be transformative, but choose not to work on it due to factors such a (perceived) lack of personal fit or opportunity, personal circumstances, or other practical considerations.

There may be various other reasons why people choose to work on other areas, despite believing transformative AI is very likely, e.g. decision-theoretic or normative/meta-normative uncertainty.

Thanks for asking ezrah. We currently plan to leave the survey open until December 31st, though it’s possible we might extend the window, as we did last time. 

I think the possibility that outreach to younger age groups[1] might be net negative is relatively neglected. That said, the two possible reasons suggested here didn't strike me as particularly conclusive.

The main reasons why I'm somewhat wary of outreach to younger ages (though there are certainly many considerations on both sides):

  • It seems quite plausible that people are less apt to adopt EA at younger ages because their thinking is 'less developed' in some relevant way that seems associated with interest in EA.
    • I think something related to but distinct from your factor (2) could also be an influence here, namely reaching out to people close to the time when they are making relevant decisions might be more effective at engaging people.
  • It also seems possible (though far from certain) that the counterfactual for many people engaged by outreach to younger age groups, is that they could have been reached by outreach targeted at a later date, i.e. many people we reach as high schoolers could simply have been reached once they were at university. 

These questions seem very uncertain, but also empirically tractable, so it's a shame that more hasn't been done to try to address them. For example, it seems relatively straightforward to compare the success rates of outreach targeting different ages. 

We previously did a little work to look at the relationship between the age when people first got involved in EA and their level of engagement. Prima facie, younger age of involvement seemed associated with higher engagement, though there's a relative dearth of people who joined EA at younger ages, making the estimates uncertain (when comparing <20s to early 20s, for example), and we'd need to spend more time on it to disentangle other possible confounds.

 

 

  1. ^

    Or it might be that 'life stages' are the relevant factor rather than age per se, i.e. a younger person who's already an undergrad might have similar outcomes when exposed to EA as a typical-age undergrad, whereas reaching out to people while in high school (regardless of age) might be associated with negative outcomes.

I think the difference between me and Yudkowsky has less to do with social effects on our speech and more to do with differing epistemic practices, i.e. about how confident one can reasonably be about the effects of poorly understood future technologies emerging in future, poorly understood circumstances. 

This isn't expressing disagreement, but I think it's also important to consider the social effects of our speaking in line with different epistemic practices, i.e.,

  • When someone says "AI will kill us all" do people understand us as expressing 100% confidence in extinction, or do they interpret it as mere hyperbole and rhetoric, and infer that what we actually mean is that AI will potentially kill us all or have other drastic effects
  • When someone says "There's a high risk AI kills us all or disempowers us" do people understand this as us expressing very high confidence that it kills us all or as saying it almost certainly won't kill us all.

I think these questions are relevant in a variety of ways:

  • Whether overall public awareness is high or low seems relevant to outreach in various ways, in different scenarios.
    • For example, this came up just a few days here in a discussion of outreach. In addition to knowing overall sentiment, knowing the overall level of awareness of EA is important, since it informs us about the importance and potential for change in sentiment (e.g., in this case, it seems very few people are even aware of EA at all, so even if negative sentiment had increased, its scope would be limited).
    • In general, after major public events pertaining to EA (like FTX), we might want to know whether these have affected awareness of EA (for good or ill), so we can respond accordingly.
    • Knowing the overall level of awareness of EA in the population (the 'top of the funnel') also informs us about the shape of the funnel, and how many people drop out after the first exposure stage, which is relevant to assessing how many people are interested in EA (as it is currently presented).
    • Still more generally, if we have any sense of what the ideal growth rate or size of EA should be (decision-makers' views on this are explored in the forthcoming results from Meta Coordination Forum Survey), then we presumably want to know where the actual growth rate or size falls relative to that.
  • Knowing about how awareness of EA varies across different groups is also relevant to our outreach.
    • For example, it could inform us about which groups we should be targeting more heavily to ensure we reach those groups.
    • It could also help identify which groups we are trying to reach but failing to make aware of EA (for whatever reason).
    • Moreover, if we know that some groups are more heavily represented in the EA community, then knowing how many people from those groups have heard of EA in the first place informs us about what point in the funnel the problem is (people not hearing about EA, hearing about it but not liking it, hearing about it, joining the community and then dropping out etc.). Our data does suggest some such disparities at the level of first-awareness for both race and gender.
  • Knowing about public sentiment towards EA seems directly relevant for outreach.
    • For example, post-FTX there was much discussion about whether the EA brand had become so toxic that we should simply abandon it (which would have entailed huge costs, even if it had been the right thing to do on balance). I won't elaborate too much on this since it seems relatively straightforward.
  • Knowing about difference in sentiment across groups is also relevant.
    • For example, if sentiment dramatically differed between men and women, or other demographics, this would potentially suggest the need for change (whether in terms of our messaging or features of the community etc.

One move which is sometimes made to suggest that these things aren't relevant, is to say that we only need to be concerned about awareness and attitudes among certain specific groups (e.g. policymakers or elite students). But even if we think that knowing about awareness and attitudes towards EA among certain groups is highly important, it doesn't suggest that broader public attitudes are not important. 

  • For example, even in cases where EA were supported by elites (of whatever kind) action may be difficult in the face of broad, public opposition.
  • The attitudes of elites (or whatever other specific, narrow group we think is of interest) and broader public opinion are not completely autonomous, so broader awareness and attitudes are likely to penetrate whatever other group we're interested in.
  • I think we actually are interested in the awareness, attitudes and involvement of a broader public, not just specific narrow groups, particularly in the long-term. At the least, some subsets of EA are interested in this, even if other subsets of EA actors might be focused more narrowly on particular groups.[1]
  1. ^

    As a practical matter, it's also worth bearing in mind that large representative surveys like this can generate estimate for some niche subgroups, for example, just not really niche ones like elite policymakers), particularly with larger sample sizes. 

We didn't directly examine why worry is increasing, across these surveys. I agree that would be an interesting thing to examine in additional work.

That said, when we asked people why they agreed or disagree with the CAIS statement, people who agreed mentioned a variety of factors including "tech experts" expressing concerns and the fact that they had seen Terminator etc., and directly observing characteristics of AI (e.g. that it seemed to be learning faster than we would be able to handle). In the CAIS statement writeup, we only examined the reasons why people disagreed (the responses tended to be more homogeneous, because many people were just saying ~ it's a serious threat), but we could potentially do further analysis of why they agreed. We'd also be interested to explore this in future work.

It's also perhaps worth noting that we originally wanted to run Pulse monthly, which would allow us to track changes in response to specific events (e.g. the releases of new LLM versions). Now we're running it quarterly (due to changes in the funding situation), that will be less feasible.

Load more