My submission is linked for the "EA Criticism and Red-Teaming Contest". Thanks for reading if you have time.
Edit started at 23:07 UTC on Thursday 1 September
N.B.: my in-text hyperlinks (i.e. the ones where the text is clickable) are not showing up on the Google Drive, but if you hover over the text as you're reading, your cursor should indicate where there is a link. It should hopefully be obvious from context where I have included such a link.
In response to the pleasant feedback in the comment section, I am now including a summary of the linked essay. The summary is formatted in terms of roughly the same sections as the essay.
Summary
1 Introduction
I argue that there is an intellectual core to Effective Altruism that I call "Economism", meaning the over-generalisation of the toolset and ideas from Neoclassical economics. I associate this with a broader way of thinking that I label the "Formal/Comprehensive Mentality". This mentality is defined by its attraction to simple formalisms and totalising frameworks. I contrast it with other mentalities that are more "integrative"---more interested in systems and emergent features---which I claim are associated with intellectual tendencies that take a very different approach than economics, such as complexity science and ecology.
2 Economism in EA: An Overview
This section focusses on creating the connection between tendencies in economics and in Effective Altruism, with little in the way of explicit criticism. I identify four areas in which the intellectual methodology of EA can be considered "economistic":
- the use of Expected Utility Theory as a universal decision criterion;
- methodological individualism and reductionism;
- the valorisation of “optimisation”;
- a belief in the power of non-empirical formal models to represent complex systems.
I consider 1. to be a core and obvious connection.
For 2., I show that economics is known for this approach to scientific methodology especially in comparison with other social sciences. I argue that EA also exhibits an individualist approach to ethical calculations in general, focussing on individual impact rather than collective responsibility, a distinguishing feature from other activist or political movements. I argue that methodological individualism is also an inherent feature of the technical apparatus of Act Consequentialism (e.g. in the study of Population Ethics), insofar as it is concerned with static aggregation of individual units. (As opposed to other approaches to ethics, which might consider these problems relatively unimportant because of a less reductionist approach.)
I deal with 3. only in brief. I connect it to expected utility maximisation (the core idea of Expected Utility Theory), and suggest it shows itself most clearly within EA in the arguments for runaway AI.
For 4., I first argue that this is a reasonable characterisation of neoclassical economics. I then argue that the clearest example of this is again the arguments for runaway AI.
3 The Flaws of Economism under Complexity
This is the main body of the essay, the section where I make my critique. There are two top level sections which each break into several sub-sections (with sub-sub-sections below).
3.1 The Breakdown of Expected Utility Theory
In the first part of this, I make an extended technical detour into the idea of "ergodicity" and how it illustrates that expected utility maximisation is not a universal algorithm for rationality.
I argue, following Ole Peters and Murray Gell-Man, that the stochastic processes we face in the everyday world, such as financial dynamics, are "non-ergodic", meaning path-dependent. An important dual-characteristic of real-world processes that guarantees this is the existence of "absorbing barriers" (as Nassim Taleb calls them) paired with multiplicative dynamics. Expected Utility Theory cannot reckon with this as it is a timeless theory of one-off maximisation of an ensemble-average of states, i.e. ergodic.
I observe that John Kelly, with his Kelly Criterion, already identified the correct policy to deal with making decisions over time. Peters and Gell-Mann, though, argue that we can turn this into a different paradigm for decision theory. One can consider this a "dynamic" decision theory to replace the "static" decision theory that is EUT.
I argue that the failure to reckon with these facts is associated with a number of flaws within the main intellectual approach of EA:
- A formal approach to ethics that focusses on static aggregation of utility
rather than consideration of dynamic processes. - Disrespect for the idea of evolutionary knowledge and “ecological rationality”.
- Disrespect for sustainability/anti-fragility in favour of optimisation.
- The weighting of highly improbable upside outcomes too highly.
- Unrealistic arguments for the threat of runaway AI.
For 1., I argue that the fundamental paradigm of Act Consequentialism, based on the framework of EUT, is a kind of "static ethics", meaning concerned with one-off maximisation of states. I oppose this with the concept of "dynamic ethics", meaning an approach to ethics that focusses on processes or policies. I make the claim: conventional Decision Theory : Peters/Gell-Mann Decision Theory :: static ethics : dynamic ethics. I claim that it is possible to find dynamic ideas within EA, such as the popular view that, as a civilisation, we should follow the policy of maximising economic growth. But I suggest that this is a simplistic policy that is misaligned with the long-term interests of the planet, and see this as a clear example of economism in action.
For 2., the headline idea is that a focus on individual optimisation blinds us to the fact that most knowledge is not the result of conscious processing, but embedded within the wiring of our brain and the structure of our society.
The "heuristics and biases" research program of Kahneman and Tversky is popular within EA circles, and sees humans as irrational in comparison to the model of Homo economicus. But the Ecological Rationality theory of Gerd Gigerenzer gives us an alternative perspective, one which sees many of the so-called flaws of human thinking as adaptations to the conditions imposed by having to make decisions in the (complex, evolving) real world. I point out that there is a natural consilience between this and Peters and Gell-Mann's dynamic decision theory.
Advocates for the idea of cultural evolution, such as the evolutionary anthropologist Joseph Henrich, have argued that many of our cultural practices and institutions are adapted to their ecological conditions. I connect this to some of the standard arguments in the Conservative tradition of political philosophy, in particular, the idea that our institutions and norms often have a vital logic that we do not understand. The moral of this is that, when calculating payoffs, we should always ask the question: How might this intervention affect the system that informed my calculation of payoffs to begin with?
For 3., I refer to Taleb's popular book Antifragility: Things that Gain from Disorder, in which one of his main ideas is that optimisation = fragility. I explain the logic of this idea by analogy with the idea of "overfitting" in Machine Learning. I point to real-world examples of where, as a society, we have become unstuck by this---for example, in the over-optimisation of global supply chains. I then make the connection to the environmental crisis that the world is experiencing. I suggest that a belief in the idea of optimisation tends to push one towards the idea that we can innovate our way out of this crisis. I quote a passage from What We Owe the Future which indicates that Will MacAskill holds such a view. I make clear my own feeling of doubt and fear about this, quoting a relevant statement from Geoffrey West: “We’re not only on a treadmill that’s going faster,
but we have to change the treadmill faster and faster” (The Surprising Math of Cities and Corporations).
For 4., I bring up another paper by Ole Peters, the Time Resolution of the St Petersburg Paradox. This holds that we can deflate the St Petersburg Paradox by realising that, if we imagine playing repeated St Petersburg gambles over time, the correct policy would be the Kelly Policy (which does not blow up to infinity). I then connect this to the ideology of "Long-termism" advocated by Nick Bostrom and MacAskill. I grant the idea that we should care about long-term stuff but not "ultra" long-term things, such as the possible lives of digital beings thousands or millions of years into the future. I suggest that the non-ergodic perspective lends support to this, because it shows us that there is an asymmetry between downside and upside: we should care a lot more about annihilation than about a tiny possibility of a great outcome.
For 5., I argue that both Nick Bostrom and Stuart Russell's arguments for the risk of runaway AI treat the world as a game, and the hypothetical AI agent as Homo economicus. I first observe their apparent belief that it's reasonable to expect a highly intelligent AI system to be a scientific and strategic genius, even though its goal system is really "dumb" compared to our complex and dynamic goal systems as humans. I tentatively suggest at a possible contradiction here. I then bring up the scenario of an AI system that has a simple goal system and then "recursively self-improves" its intelligence while maintaining that goal system. I argue that this is implausible for 3 reasons, which I will simplify into 2 for this summary:
(1.) The possibility of recursive self-improvement sneaking up on us is highly questionable, given that the necessary starting point for such a process is an agent that has a conception of self and knows how the world works to a large extent.
(2.) The way in which we learn, as humans, to thrive in multiple, complex environments, is by having a complex goal system that pushes us to explore multiple, complex environments as we are learning, so why wouldn't this be our default assumption for an AGI?
I argue a core belief behind Bostrom and Russell's arguments is the idea that the world is little more than a game that an agent smart enough can control. But I observe that no agent could control the world entirely, and suggest that it is therefore rational to expect a highly intelligent agent, artificial or not, to want to explore as well as exploit.
3.2 False Precision and Aggregation
This section is shorter and breaks into two main sub-sections.
3.2.1 Bayesianism and its Scope
In this section, I define Bayesianism as "an ideology that advocates the rationality of trying to assign subjective probabilities to beliefs about ordinary things". I argue that this ideology has a number of flaws.
There is a typo where I say I have identified "four flaws" but there are in fact 5:
- The Feasability/Authenticity Problem
- The Introspection/Realism Problem
- The Uncertainty Problem
- The False Precision Problem
- The Neglect of Imprecision Problem
For 1, I argue that, under one interpretation, Bayesianism requires superhuman cognitive prowess. Using Kahenmann's System 1/System 2 distinction, a 'true' Bayesian needs to have System 2 switched on all the time. But this is impossible.
For 2, I start by casting doubt on one of the foundational ideas of cognitive science, Jerry Fodor's conception of ‘Folk Psychology’, which posits that the human mind is a web of “beliefs”. I suggest that this is a language-centric view of things and offer an alternative view, pithily summarised by Yann LeCunn in a tweet: "Language is an imperfect, incomplete, and low-bandwidth serialization protocol for the internal
data structures we call thoughts." Instead of being driven by "beliefs", I posit that thought is driven by "world models". Like most mathematical models in science, these world models may possibly predict in a format we could compare to a probability distribution but not single-point probabilities. I propose this as an explanation for why it is so un-intuitive to assign subjective probabilities to one-off events: because it is not thought-native.
For 3, I refer to the idea of "fundamental uncertainty", as advocated by John Maynard Keynes, that many problems of prediction are so complex (or ill-defined) that it is simply inappropriate to apply probability theory to them. I suggest that this reluctance to assign probabilities to beliefs about specific events is the default view of people in part because they know that they don't have confidence in their beliefs. I grant that there is an extra accountability that comes with the attitude of assigning probabilities and especially betting on these probabilities. But I point out that an equally accountable alternative is just owning up to one's uncertainty.
For 4, I make reference to the prediction aggregation site Metaculus. I point out that the methodology of predicting in terms of single-point probabilities gives inherently favourable-looking results compared to, say, the way in which we test scientific models that, by modelling an entire system, imply an entire probability distribution. I point to the fact that the Metaculus community has a decent-looking Brier score but also has gotten quite a few thing completely wrong.
For 5, I discuss the idea that the practice of trying to make precise predictions inherently pushes one into the realm of problems where that practice isn't inherently doomed. I refer to Phillip Tetlock's implicit acknowledgement of this idea in Superforecasting. One important class of phenomena that super-forecasters knowingly can't predict is Black Swan Events, phase changes in chaotic systems that can't be timed like financial cascades, earthquakes or (arguably) pandemics. I observe that while Black Swans can't be timed, there are typically structural indicators of hazard that people can observe and warn about. Since these warnings will not correspond to precise quantitative predictions, the Bayesian mindset may be inclined to dismiss them. From here, I make the broader point that there is plenty of knowledge and information in non-quantitative fields in general, it's just a little 'chewier'.
I finish by pointing out that a focus on single-point probabilities also tends to cause a neglect of processes that have spectral outcomes. I argue that Toby Ord's focus in The Precipice of the single-point probability of existential risk causes him to believe that it is more urgent to address AI risk than climate change, even though we know that every possible climate change outcome in the spectrum is awful, even though many AI outcomes are possibly quite good, even on his account.
3.2.1 Belief Aggregation and the Efficient Markets Hypothesis
This is a much shorter section than the previous. A common defence of prediction aggregation systems like Metaculus, or prediction markets---both of which are popular concepts in EA---refers to the idea of market efficiency. I use this section simply to suggest that this defence is a little weaker than it at first appears.
In the first part of this, I interrogate the meaning of the Efficient Markets idea. I note that an efficient market is usually defined as one where security prices “reflect all available information”, which is sometimes held to carry the idea that security prices reflect "reality", in some sense. But I note that the tests developed for empirically validating market efficiency test for something that is arguably a little different: whether traders can non-randomly beat the market. I claim that security prices do rapidly gobble up all/almost all of the "information" that traders have, but that this information is often just information about belief. Even market bubbles, then, can be efficient, in the sense that nobody can really work out when the reservoir of belief will run out. If everyone in the market was a rational agent, a la Homo economicus, this wouldn't be possible, because the Nash equilibrium is not to invest in the bubble stock, if there is common knowledge among investors that the stock could plausibly lose all revenue. But we are not, and therefore we shouldn't necessarily think that security prices are directly hooking onto facts about the world outside the market (or outside the realm of belief itself).
I don't make too much of this point, only suggest that it casts doubt on "the idea that lots of people trying to make guesses at a single value will necessarily make that value more aligned with the facts of the world".
Conclusion
I connect all of the above to the idea of a "mentality". I finally define the concept of the "Formal/Comprehensive Mentality". I say the "Formal" part is self-explanatory. The “Comprehensive” part "captures the idea of a mindset that believes intellectual topics can be completely wrangled—isolated and conquered—with one master approach or methodology."
I suggest there is at least one clear benefit to this mentality, which is that the reification of simple models is very helpful for motivating people to action. This is at the heart of EA as a movement. It can be seen in the way in which MacAskill and Ord took Singer's thought in "Famine, Affluence and Morality" and ran with it. But another core idea of EA is that knowledge should precede action, and I argue that some of the intellectual methodologies of the movement cannot service that goal.
I note that my own way of thinking is different from this, which leads me away from the ideas of economism and towards ideas in e.g. complexity science, ecology and cultural evolution. I speculate that the identity of EA as a movement may inherently correlate with certain mentalities but claim that there are things that EA could learn from alternate ways of thinking, if it could find a way to integrate them.
I found this to be a comprehensive critique of some of the EA community's theoretical tendencies (over-reliance on formalisms, false precision, and excessive faith in aggregation). +1 to Michael Townsend's suggestions, especially adding a TLDR to this post.