The Effective Altruism website defines EA as: "We use evidence and careful analysis to find the very best causes to work on." The Introduction to Effective Altruism post in our forum also says: "It is a research field which uses high-quality evidence and careful reasoning to work out how to help others as much as possible."
So I guess this is more or less considered the definition of EA. But as I read more about EA, I am beginning to feel like this definition may be insufficient. It looks like the EA focus splits across two schools of thought - Evidence-based giving and hits-based giving. But this definition seems like it is all about Evidence-based giving. It feels like the 'GiveWell-ness' of it all is represented but what about the 'OpenPhil-ness'?
This exclusion of hits-based giving from the definition seems problematic since 80000hours.org (one of the top 5 ways through which people actually find EA) considers Expected Value thinking (the foundation of Hits Based giving if I understand it correctly) as one of the key ideas of EA. But then you see the definition and it is not really there. In addition, the incompleteness of the definition could also make it difficult for someone to see why EA does GCR work, in my opinion. Please correct me if I am wrong but it feels like GCRs doesn't necessarily have high-quality evidence for why we should work on it but Expected Value thinking is what really makes it worth it.
UPDATE:
I had only mentioned two sources of definitions above. But there could be more that I may have missed. If you know of more please mention them in the comments/answers and I will add them to this list:
- Defining Effective Altruism by William_MacAskill. Thanks to Davidmanheim for bringing this up in the answer here. The definition given in Will's post is:
Effective altruism is: (i) the use of evidence and careful reasoning to work out how to maximize the good with a given unit of resources, tentatively understanding ‘the good’ in impartial welfarist terms, and (ii) the use of the findings from (i) to try to improve the world.
One, I'd argue that hits-based giving is a natural consequence of working through what using "high-quality evidence and careful reasoning to work out how to help others as much as possible" reallying means, since that statement doesn't say anything about excluding high-variance strategies. For example, many would say there's high-quality evidence about AI risk, lots of careful reasoning has been done to assess its impact on the long term future, and many have concluded that working on such things is likely to help others as much as possible, though we may not be able to measure that help for a long time and we may make mistakes.
Two, it's likely a strategic choice to not be in-your-face about high variance giving strategies since they are pretty weird to most people. EA orgs have chosen to develop a public brand that is broadly appealing and not controversial on the surface (even if EA ends up courting controversy anyway because of its consequences for opportunities we judge to be relatively less effective than others). The definitions of EA you point to seem in line with this.
Agreed - see my answer which notes that Will suggested a phrasing that omits "high-quality."