Hide table of contents

Systems engineers, project managers and other professionals have a “Body of Knowledge,” the set of things that all members of the profession are supposed to know. Likewise, I argue that there is a minimum set of knowledge that allows people to understand the debates in the effective altruist community and participate effectively in its discussions. I don’t mean domain knowledge in AI alignment or poverty alleviation, but rather what is needed to evaluate and prioritize issues and proposals.

So, if I am right, what should this Effective Altruist Body of Knowledge (EALBOK) include? And should we encourage effective altruists to make it a part of their higher education? The following list is incomplete; it doubtless reflects my expertise, my interests, and my prejudices. But I believe it is a good starting point.


THE BASICS

· Economics. Effective altruists should understand utility, trade-offs, incentives, and discounting. All this should be covered in your standard beginner microeconomics course. Or study it for free here: https://www.edx.org/course/microeconomics. Paradoxically for people who think big, the study of macroeconomics would likely be much less useful.

· Probability and statistics. Probability deals with predicting the likelihood of future events, while statistics involves the analysis of the frequency of past events. Concepts such as conditional probability, the main probabilistic distributions, and correlation are necessary to understand much of the effective altruist literature in any depth. I would also argue that they are essential for any educated person. There are many free online courses in probability, and even more in statistics; pick your poison: https://www.my-mooc.com/en/categorie/statistics-and-probability.

· Decision analysis. The whole point of effective altruism is to address important decisions in a formal manner, taking into account our values (i.e., our utility function), our time and risk preferences, and uncertainty (represented by probability distributions, most of them subjective). Effective altruists should consider how this “ideal” prescriptive model is affected by biases. Bonus points for understanding how Monte Carlo simulation and discretization work. Decision analysis courses are uncommon, especially at the undergraduate level, but you can find much of the material here: https://online-learning.tudelft.nl/courses/effective-decision-making-dealing-with-business-complexity/. If you want to study the matter in depth, there is an excellent free textbook at https://smartorg.com/wp-content/uploads/2011/08/Decision-Analysis-for-the-Professional.pdf. Decision analysis is tremendously useful in everyday life, whether you are an altruist or a misanthropist.

· Big history. Bill Gates strongly endorsed the teaching of big history, and I concur. Big history studies the history of our Universe, from the Big Bang to date, searching for universal patterns and trends. It provides much needed “situational awareness” for effective altruists. It is interesting, eye-opening, and free at https://www.coursera.org/learn/big-history.


STAGE TWO

· Game theory. Best known for its use in economics, game theory underlies concepts as relevant to effective altruists as the tragedy of the commons. You can do worse than the introductory course at https://www.coursera.org/learn/game-theory-1. For the serious reader, a comprehensive but accessible free textbook is available at http://faculty.econ.ucdavis.edu/faculty/bonanno/PDF/GT_book.pdf.

· System dynamics simulation. System dynamics is a technique for understanding socio-technical systems over time using mathematical models. What makes system dynamics different from other approaches is the use of feedback loops and stocks and flows. These elements help describe how even seemingly simple systems produce complex behavior. I am probably biased, but I find system dynamics tremendously useful for understanding phenomena such as the spread of epidemics and climate change. While geared towards medical applications, the course at https://www.edx.org/course/system-dynamics-for-health-sciences should cover the basics. If you like to do your own modeling, I recommend the Vensim package. Its Personal Learning Edition has good documentation and can be downloaded for free at https://vensim.com/free-download/.

· Expert judgment. Many of the inputs for the models described above are “expert judgments,” i.e., best guesses by knowledgeable people. How to elicit (i.e., get the experts to talk) and aggregate (i.e., average the experts when they disagree) this often tacit knowledge is a growing field, and the free course at https://www.my-mooc.com/en/mooc/decision-making-under-uncertainty-introduction-to-structured-expert-judgment/ should cover the basics. Philip Tetlock has distilled expert judgment into a science, identifying “superforecasters” who routinely outperform leading experts in their fields. Check out Tetlock’s book at https://www.amazon.com/Superforecasting-Science-Prediction-Philip-Tetlock-ebook/dp/B00Y78X7HY (it’s cheap).

· Long-range planning. The RAND Corporation is probably the gold standard in long-range planning under uncertainty. People who obsess about the future will thus be interested in the recent textbook, Decision Making under Deep Uncertainty. This is cutting-edge stuff, so do not expect a free online course anytime soon. But the free textbook can be found at https://link.springer.com/content/pdf/10.1007%2F978-3-030-05252-2.pdf.


A big absence here is that of any mathematics beyond high school algebra. While calculus can make the above concepts much more powerful, it often introduces an algebraic barrier that prevents true understanding. If you have, or want, a calculus background, go for it. But you can survive without it.

An even more glaring omission is that of philosophy, especially ethics. What should we want? Our values are not “just there.” However, I have preferred not to include Philosophy in the above list for four reasons:

1. Much useful work can be done using the simple utilitarian framework that is tacit, for instance, in economics.

2. I find it difficult to offer prescriptive ethical advice. To my knowledge, there is no “right set” of values; at least not yet.

3. I suspect that our values are not arrived at rationally, but gradually and holistically. How much does the study of philosophy impact our values? I don’t know if there is any data out there; my uneducated guess is “not much.”

4. I cannot claim any particular expertise in philosophy, so I will defer to the advice of those better qualified.

48

0
0

Reactions

0
0

More posts like this

Comments11


Sorted by Click to highlight new comments since:

Hmm. On the one hand I think these are all useful topics for an EA to know. But I don't think it's necessary for all EAs to know these things. I think there's a lot of EAs who don't have this technical knowledge, but are happy to outsource decisions relying on this knowledge (such as where to donate) to people who do. That said, I think that often leads to donating less-than-effectively (e.g. giving to whatever EA Fund appeals to you personally, rather than rationally thinking about trade-offs/probabilistic outcomes).

I guess this is, in part, a big-tent vs. elite EA trade-off question. If EA is best as an elite movement, it makes sense that all the members should have this knowledge. But if we want to take an "everyone has a place in EA" approach, then it might not make sense to have a central curriculum.

Also, I don't think we want everyone in EA to have the same skillset. EA isn't, in my view, a single professional field, but perhaps more like a company (although this is probably an oversimplification). If a company gave all of their employees a handbook on How to Be A Great Project Manager, it'd be helpful... for project managers. But the rest of the team ought to be rounding out skills that others in the company don't have that suit their comparative advantage and will move the company forward. The only thing everyone at the company really needs to know is the product. Basic time management / other soft skills are also useful. I don't think we need 100% of EAs to have a solid grounding in economics. Maybe we need ~100% of EAs to trust economics. But I'd rather have some EAs focusing on building skills like movement-building, communications, fundraising, operations/management, entrepreneurship, policy, qualitative research, etc.

Granted, I'm thinking about this from the perspective of careers, rather than being able to participate in discussions in EA spaces. To answer to that aspect of it - although I certainly think it's possible to discuss EA without knowing about economics / statistics / decision analysis knowledge, the conversation does sometimes go in this more technical direction and leave newcomers behind. The question, then, might be whether it's the newcomers who should hold the responsibility of learning this so that they can participate in these discussions, or if the people who are discussing things at such a technical level should adjust the way they discuss these issues to make them more accessible to a non-technical audience. I lean more towards the latter (though it depends on the context).

[anonymous]5
0
0

I agree with Marisa

Rather than a single body of knowledge being a standard education for EAs, I like the fellowship structure that many EA Uni groups use.

For me, one of the main goals in running these fellowship to expose students to enough EA ideas and discussions to decide for themselves what knowledge and skills they want to build up in order to do good better. For some people, this will involve economics, statistics, and decision analysis knowledge, but for others, it will look totally different.

(For fellowship syllabus examples you can check out this Intro Fellowship I'm running at Brown EA, and this In-Depth Fellowship run by EA Oxford).

I do think some ethics is a must, not necessarily to be prescriptive, but to challenge people's views and introduce alternatives so they don't get stuck with something they would not endorse if they knew more. Some topics I'd recommend:

  • Population axiology: total utilitarianism, the repugnant conclusion (and other similar results), person-affecting views, negative utilitarianism, average utilitarianism, prioritarianism, egalitarianism, lives vs headaches, dust specks vs torture. Hilary Greaves wrote a survey on population axiology.
  • Welfare/wellbeing: hedonistic, desire-based, objective list theory. Symmetric vs asymmetric/antifrustrationist/negative/suffering-focused views. See the SEP article.
  • Impartiality, just and unjust discrimination, equal consideration of interests, speciesism. Moral subjects/patients vs moral agents, sentience/consciousness.
  • Free will, personal identity, death, moral luck, moral responsibility.
  • Objections and alternatives to consequentialism.
  • Metaethics, realism vs anti-realism, moral uncertainty.

The Stanford Encyclopedia of Philosophy articles are usually good, and sometimes the Wikipedia articles are, too.

This stuff is interesting to think about. There have been EA courses before. There could one-day be a textbook for effective altruism. There could be a successor to the RSP that offers a degree. Similar stories for "global prioritisation", "macrostrategy", and "AI safety".

Julia Wise provided a list of EA syllabi and teaching materials here.

(RSP = Future of Humanity Institute's Research Scholars Programme.)

There are also the reading lists recently put together by Richard Ngo.

Great question! I hope to find time to engage substantively later, but for now I just wanted to flag that I'm considering to spend significant time from September or October putting together some kind of "EA curriculum", and that I'd be happy to talk to anyone interested in similar ideas. Send me a PM if you want to jump on a call in the next couple of weeks.

[anonymous]3
0
0

Hi Max,

I'm curious how big you are thinking this "EA curriculum" might be. Are you thinking of something similar to an EA Uni group fellowship (usually ~4 hours/ week for ~ 8 weeks) or are you thinking of something much larger?

I was mostly thinking of a curriculum that would eventually be much larger (though could be modular, and certainly would have a smaller MVP as first step to gauge viability of the larger curriculum).

But my views on this aren't firm, and in general one of the first things I'll do is to determine various fundamental properties I don't feel certain about yet. Other than length these are e.g. target audience and intended outcomes (e.g. attracting new people to EA, "onboarding" new EAs, bringing moderately experienced EA to the same level, or allowing even quite involved EAs to learn something new by increasing the amount of content that publicly accessible as opposed to in some people's minds or nonpublic docs), scope (e.g. only longtermism?), and focus on content/knowledge vs. skills/methods.

[anonymous]1
0
0

Sounds very exciting!

And seems like there is some overlap with EA Uni group fellowships so I would be happy to talk to you about those if you want; although maybe better to talk to the community builders more involved in syllabus writing. ( this Intro Fellowship I'm running at Brown EA )

I remember commenting to an economist friend a few months ago that economists have generally much better ethics than philosophers, precisely because of their consistent application of utilitarianism followed by moving on to the interesting questions, as opposed to philosophers wanting to debate ethics to death. So I concur with the decision to include economics over philosophy.

Curated and popular this week
 ·  · 25m read
 · 
Epistemic status: This post — the result of a loosely timeboxed ~2-day sprint[1] — is more like “research notes with rough takes” than “report with solid answers.” You should interpret the things we say as best guesses, and not give them much more weight than that. Summary There’s been some discussion of what “transformative AI may arrive soon” might mean for animal advocates. After a very shallow review, we’ve tentatively concluded that radical changes to the animal welfare (AW) field are not yet warranted. In particular: * Some ideas in this space seem fairly promising, but in the “maybe a researcher should look into this” stage, rather than “shovel-ready” * We’re skeptical of the case for most speculative “TAI<>AW” projects * We think the most common version of this argument underrates how radically weird post-“transformative”-AI worlds would be, and how much this harms our ability to predict the longer-run effects of interventions available to us today. Without specific reasons to believe that an intervention is especially robust,[2] we think it’s best to discount its expected value to ~zero. Here’s a brief overview of our (tentative!) actionable takes on this question[3]: ✅ Some things we recommend❌ Some things we don’t recommend * Dedicating some amount of (ongoing) attention to the possibility of “AW lock ins”[4]  * Pursuing other exploratory research on what transformative AI might mean for animals & how to help (we’re unconvinced by most existing proposals, but many of these ideas have received <1 month of research effort from everyone in the space combined — it would be unsurprising if even just a few months of effort turned up better ideas) * Investing in highly “flexible” capacity for advancing animal interests in AI-transformed worlds * Trying to use AI for near-term animal welfare work, and fundraising from donors who have invested in AI * Heavily discounting “normal” interventions that take 10+ years to help animals * “Rowing” on na
 ·  · 3m read
 · 
About the program Hi! We’re Chana and Aric, from the new 80,000 Hours video program. For over a decade, 80,000 Hours has been talking about the world’s most pressing problems in newsletters, articles and many extremely lengthy podcasts. But today’s world calls for video, so we’ve started a video program[1], and we’re so excited to tell you about it! 80,000 Hours is launching AI in Context, a new YouTube channel hosted by Aric Floyd. Together with associated Instagram and TikTok accounts, the channel will aim to inform, entertain, and energize with a mix of long and shortform videos about the risks of transformative AI, and what people can do about them. [Chana has also been experimenting with making shortform videos, which you can check out here; we’re still deciding on what form her content creation will take] We hope to bring our own personalities and perspectives on these issues, alongside humor, earnestness, and nuance. We want to help people make sense of the world we're in and think about what role they might play in the upcoming years of potentially rapid change. Our first long-form video For our first long-form video, we decided to explore AI Futures Project’s AI 2027 scenario (which has been widely discussed on the Forum). It combines quantitative forecasting and storytelling to depict a possible future that might include human extinction, or in a better outcome, “merely” an unprecedented concentration of power. Why? We wanted to start our new channel with a compelling story that viewers can sink their teeth into, and that a wide audience would have reason to watch, even if they don’t yet know who we are or trust our viewpoints yet. (We think a video about “Why AI might pose an existential risk”, for example, might depend more on pre-existing trust to succeed.) We also saw this as an opportunity to tell the world about the ideas and people that have for years been anticipating the progress and dangers of AI (that’s many of you!), and invite the br
 ·  · 3m read
 · 
Hi all, This is a one time cross-post from my substack. If you like it, you can subscribe to the substack at tobiasleenaert.substack.com. Thanks Gaslit by humanity After twenty-five years in the animal liberation movement, I’m still looking for ways to make people see. I’ve given countless talks, co-founded organizations, written numerous articles and cited hundreds of statistics to thousands of people. And yet, most days, I know none of this will do what I hope: open their eyes to the immensity of animal suffering. Sometimes I feel obsessed with finding the ultimate way to make people understand and care. This obsession is about stopping the horror, but it’s also about something else, something harder to put into words: sometimes the suffering feels so enormous that I start doubting my own perception - especially because others don’t seem to see it. It’s as if I am being gaslit by humanity, with its quiet, constant suggestion that I must be overreacting, because no one else seems alarmed. “I must be mad” Some quotes from the book The Lives of Animals, by South African writer and Nobel laureate J.M. Coetzee, may help illustrate this feeling. In his novella, Coetzee speaks through a female vegetarian protagonist named Elisabeth Costello. We see her wrestle with questions of suffering, guilt and responsibility. At one point, Elisabeth makes the following internal observation about her family’s consumption of animal products: “I seem to move around perfectly easily among people, to have perfectly normal relations with them. Is it possible, I ask myself, that all of them are participants in a crime of stupefying proportions? Am I fantasizing it all? I must be mad!” Elisabeth wonders: can something be a crime if billions are participating in it? She goes back and forth on this. On the one hand she can’t not see what she is seeing: “Yet every day I see the evidences. The very people I suspect produce the evidence, exhibit it, offer it to me. Corpses. Fragments of