finm

Researcher @ Longview Philanthropy
3142 karmaJoined Working (0-5 years)Oxford, UK
www.finmoorhouse.com/writing

Bio

I do research at Longview Philanthropy. Previously I was a Research scholar at FHI and assistant to Toby Ord. Philosophy at Cambridge before that.

I also do a podcast about EA called Hear This Idea.

www.finmoorhouse.com/writing

www.hearthisidea.com

Comments
168

Yeah, fair! I guess there's a broad understanding of utilitarianism, which is “the sum of any monotone or non-decreasing transformation of utilities", and a narrower understanding, which is “the sum of utilities”. But I want to say that prioritarianism (a version of the former) is an alternative to utilitarianism, not a variant. Not actually sure what prioritarians would say. Also not really an important point to argue about.

Glad to have highlighted the cardinality point!

And that meant we could hammer out more of the details in the comments, which seems appropriate.

Agree! As I say, I feel much clearer now on your position.

Thanks for all this! I agree that something like Nash is appealing for a bunch of reasons. Not least because it's Pareto efficient, so doesn't screw people over for the greater good, which feels more politically legitimate. It is also principled, in that it doesn't require some social planner to decide how to weigh people's preferences or wellbeing.

My sense, though you know much more about all this, is that Nash bargaining is not well described as a variant of utilitarianism, though I case it's a grey area.

Maybe I'm realising now that a lot of the action in your argument is not in arguing for the values which guide the future to be democratically chosen, but rather in thikning through which kinds of democratic mechanisms are best. Where plain old majority rule seems very unappealing, but more granular approaches which give more weight to those who care most about a given issue look much better. And (here we agree) this is especially important if you think that the wrong kind of popular future, such as a homogenous majority-determined future, could fall far short of the best future.

I really am thinking this as more like "what values should we be aiming at to guide the future" and being fairly agnostic on mechanism

Thanks, super useful! And makes me much more clear on the argument / sympathetic to it.

Of course, all the labs say things about democratic inputs into AI model specs and are trying stuff to this end, and these could be deeper or shallow.

Ah, that's a good point! And then I guess both of as are in some kind of agreement that this kind of stuff (deliberate structured initatives to inject some democracy into the models) ends up majorly determining outcomes from AGI.

>Re: what deep democracy, I like Nash bargaining!

I find this somewhat confusing, because elsewhere you say the kind of deep democracy you are imagining "is basically equivalent to enlightened preference utilitarianism". But the Nash solution is basically never the utility-maximising solution! You can even set things up so that the Nash solution is an arbitrarily small fraction, in terms of total welfare, between the disagreement point and the utility-maxising optimum. I do think there is an interesting and important question of how good, in practice, the Nash outcome is compared to the utility-maximing outcome, maybe it's great ¯\_(ツ)_/¯

Thanks for the post! Sure seems like an important topic. Unnecessarily grumpy thoughts to follow —

First, I feel unsure what you're imagining this ‘deep democracy’ thing is an answer to. You write, “If you want the long-term future to go well by the lights of a certain value function[…]” Who is “you” here?

An individual? If so, it seems clear to me that the best way to bring about things I care about, holding fixed others' strategies, is indeed to get my AI(s) to pursue things I care about, and clearly not best or desirable from any almost individual's point of view (certainly not an equilibrium) to get their AI(s) to pursue some kind of democratic conglomeration of everybody's interests.

An AI company? If so, I imagine (at least by default) that the companies are just trying to serve demand with more and better models, complying with regulation, and maybe proactively adding in side constraints and various revenue-sacrificing behaviours for prosocial or reputational reasons. It's not obvious to me that there comes a point where the executives sit down to discuss: "what set of values should we align AI to?", where the candidate answers are various (pseudo-)moral views.

In the case where we're considering a company considering the target of alignment, I wonder if my confusion comes from starting from different assumptions about how 'unitary' their product is. One outcome, which I currently view as a default, is just an extrapolation of the status quo today: companies train base models, then those base models are post-trained and broken up into various specialised variants, which can in turn be fine-tuned for specific use cases, and all eventually served for a million different uses. On this picture, it's not clear how much influence an AI company can have on the moral vision of the models they ultimately serve. The main reason is just that the vast majority of what these AIs are doing, in my mind, are just helpful or economically useful tasks based on (direct or delegated) specific human instructions, not following a grand impartial moral plan. And if the models are too eager to break ranks and pursue an abstract moral vision, people won't use them. This is what runs through my head when people talk about "the AGI" — what is that?

Of course, there are some reasons for thinking the AI landscape will be more unitary, and this picture could be wrong. Maybe a corporate monopoly, maybe a centralised (state-led) project, maybe a coup. Let's consider the extreme case where "you" are a lab exec, you hold total power over the world through some single alignable AI system, and you face the decision of what to tell it to do. Here I'd zoom in on the part where you say it would be "uncooperative, undemocratic, coercive" to implement your values. One rejoinder here is to make the point that, AI aside, you should (i) be morally uncertain, (ii) interested in figuring out what's good through deliberation, and (iii) care about other people. So if the hegemon-leader had a reasonable moral view, directly implementing it through an AI hegemon doesn't strike me as obviously worse in expectation than ceding influence. If the hegemon-leader has a moral view which is obviously bad, then I don't think it's very interesting that a democratic thing seems better.

In any case, I agree that the main reasons against the hegemon-leader directly implementing their vision of what's good. But (at a gut level) this is, as you say, because it would be illegitimate, uncooperative, etc., not to mention practically likely to fail (we haven't tested global totalitarianism, but most new political experiments fail). And I think democractic-but-hegemonic AI probably fares pretty badly by those lights, too, compared to actually just ceding power or not becoming a hegemon in the first place?

I do feel like I'm being unfair or missing something here. Maybe another reading is: look, anarchy is bad, especially when everyone has an army of AIs to carry out their bidding, and everyone is rich enough to start caring about scary, scope-sensitive, ideologically-motivated outcomes. The result is a bunch of winner-take-all conflict, and general destructive competition. So we need some governance system (national and/or global?) which curbs this destruction, but also does the best job possible aggregating people's values. And this is what "deep democracy" should be doing.

The part of this I agree with is that, as far as we have voting systems, they could be much improved post-AGI, in a million ways. Thumbs up to people imagining what those tools and approaches could look like, to make democratic political procedures more flexible, effective, rich in information, great at finding win-win compromises, and so on.

But there's a part that feels underspecified, and a part that I'm more sceptical of. The part that feels underspecified is what "deep democracy" actually is. The way you've phrased it, and I'm being a bit unfair here, is close to being good by definition ("deeply capturing and being responsive to every single person's values" — I mean, sure!) I expect this is one of those cases where, once forced to actually specify the system, you make salient the fact that any particular system has to make tradeoffs (cf Arrow, though that's a bit overblown).

The part I'm more sceptical of is that the anarchic alternative is chaotic and destructive, and the best way to aggregate preferencesis via setting up some centralised monopoly on force, and figuring out what centralised process it follows. Consider the international ~anarchy. War is a really unattractive option, even for neighboring expansionary states, so in theory (and often in practice) compromise is virtually always preferred to war. And that's the hard (fully anarchic) case — smaller-scale conflict is avoided because of criminal and civil laws which make it very not worth it.

Finally, I'd suggest that, in a sense, we have a way to allocate resources and efforts towards what people want in a granular, deep, preference-aggregating way: trade. My sense is to think about this as (even today) the main means by which society is arranged to make everyone better-off; and then consider cases where centralised processes (like voting) are necessary or valuable. One example, which seems potentially very important, is if something like "moral worth" becomes even more divorced from wealth. Of course in designing a democratic process you don't have perfectly neutral ground to stand on; you have to make a call on who gets to partake. But you can give more of a voice to people who are otherwise practically totally disenfranchised because they lack resources; while the ultra-wealthy otherwise would dominate outcomes. That's already an issue and could become a bigger issue, but does suggest an alternative answer to the question you're asking, which is (potentially a lot of) wealth redistribution.

If value is a power law or similarly distributed, then you have a high chance of at least capturing some of the stuff that is astronomically more valuable than everything else, rather than losing out on this stuff entirely.

Of course there are a bunch of methods in social choice which do this, like quadratic voting; though it's notable that most electoral democracies are not good examples, and anything like "go with the majority while protecting the rights of the minority" seems apt to highly underrate cases where some voters think that a particular issue is astronimically higher in stakes than others think. But this is also a case where I don't think there's an especially neutral, non-theory-laden approach for how to recognise how and when to give people's views more weight because (intuitively) they think some issue is higher stakes. Then again, I think this is a general problem, not a specific issue with designing democratic methods.

Ok sorry for the chaotically written thoughts, I think they probably understate how much I'm a fan of this line of thinking. And your comments in reply to this were clarifying and made me realise I was in fact a bit confused on a couple points. Thanks again for writing.

as people get richer and happier and wiser and so forth, they just have more time and interest and mental/emotional capacity to think carefully about ethics and act accordingly

I think that's right as far as it goes, but it's worth considering the limiting behaviour with wealth, which is (as the ultra-wealthy show) presumably that the activity of thinking about ethics still competes with other “leisure time” activities, and behaviour is still sensitive to non-moral incentives like through social competition. But also note that the ways in which it becomes cheaper to help others as society gets richer are going to tend to be the ways in which it becomes cheaper for others to help themselves (or ways in which people's lives just get better without much altruism). That's not always true, like in the case of animals.

Good point about persuasion. I guess one way of saying that back, is that (i) if the "right" or just "less bad" moral views are on average the most persuasive views, and (ii) at least some people are generating them, then they will win out. One worry is that (i) isn't true, because other bad views are more memetically fit, even in a society of people with access to very good abstract reasoning abilities.

Thanks Oscar!

I believe this is correct, but possibly for the wrong reason.

The reason you give sounds right — for certain concave views, you really care that at least some people are bringing about or acting on the things you care about. One point, though, is that (as we discuss in the previous post) reasonable views which are concave in the good direction are going to be more sensitive (higher magnitude or less concave) in the negative direction. If you have such a view, you might also think that there are particular ways to bring about a lot of disvalue, so the expected quantity of bads could scale similarly to goods with the number of actors.

Another possibility is that you don't need to have a concave view to value increasing the number of actors trading with one another. If there are very few actors, the amount by which a given actor can multiply the value they can achieve by their own (even linear) lights, compared to the no-trade case, may be lower, because they have more opporunities for trade. But I haven't thought this through and it could be wrong.

The choice of two dimensions is unmotivated

Yep, totally arbitrary choice to suit the diagram. I've made a note to clarify that! Agree it generalises, and indeed people diverge more and more quickly on average in higher dimensions.

Also note 4.2. in ‘How to Make the Future Better’ (and footnote 32) —

[O]n this model, making each factor more correlated can dramatically improve the expected value of the future — without improving the expected value of any individual factor at all.

Which could look like “averting a catstrophic disruption of an otherwise convergent win”.

Thanks for writing this James, I think you raise good points, and for the most part I think you're accurately representing what we say in the piece. Some scattered thoughts —

I agree that you can't blithely assume scaling trends will continue, or that scaling “laws” will continue to hold. I don't think both assumptions are entirely unmotivated, because the trends have already spanned many orders of magnitude. E.g. pretraining compute has spanned ≈ 1e8 OOMs of FLOP since the introduction of the transformer, and scaling “laws” for given measures of performance hold up across almost as many OOMs, experimentally and practically. Presumably the trends have to break down eventually, but if all you knew were the trends so far, it seems reasonable to spread your guesses over many future OOMs.

Of course we know more than big-picture trends. As you say, GPT 4.5 seemed pretty disappointing, on benchmarks and just qualitatively. People at labs are making noises about pretraining scaling as we know it slowing down. I do think short timelines[1] look less likely than they did 6 months or a year ago. Progress is often made by a series of overlapping s-curves, and the source of future gains could come from elsewhere, like inference scaling. But we don't give specific reasons to expect that in the piece.

On the idea of a “software-only intelligence explosion”, this piece by Daniel Eth and Tom Davidson gives more detailed considerations which are specific to ML R&D. It's not totally obvious to me why you should expect returns to diminish faster in broad vs narrow domains. In very narrow domains (e.g. checkers-playing programs), you might intuitively think that the ‘well’ of creative ways to improve performance is shallower, so further improvements beyond the most obvious are less useful. That said, I'm not strongly sold on the “software-only intelligence explosion” idea; I think it's not at all guaranteed (maybe I'm 40% on it?) but worth having on the radar.

I also agree that benchmarks don't (necessarily) generalise well to real-world tasks. In fact I think this is one of the cruxes between people who expect AI progress to quickly transform the world in various ways. I do think the cognitive capabilities of frontier LLMs can meaningfully be described as “broad” in the sense that: if I tried to think of a long list of tasks and questions which span an intuitively broad range of “cognitive skills” before knowing about the skill distribution of LLMs[2] then I expect the best LLMs to perform well across the large majority of those tasks and questions.

But I agree there is a gap between impressive benchmark scores and real-world usefulness. Partly there's a kind of o-ring dynamic, where the speed of progress is gated by the slowest-moving critical processes — speeding up tasks that normally take 20% of a researcher's time by 1,000x won't speed them up by more than 25% if the other tasks can't be substituted. Partly there are specific competences which AI models don't have, like reliably carrying out long-range tasks. Partly (and relatedly) there's an issue of access to implicit knowledge and other bespoke training data (like examples of expert practitioners carrying out long time-horizon tasks) which are scarce in the public internet training data. And so on.

Note that (unless we badly misphrased something) we are not arguing for “AGI by 2030” in the piece you are discussing. As always, it depends on what “AGI” means, but if it means a world where tech progress is advancing across the board at 10x rates, or a time where there is basically no cognitive skill where any human still outperforms some AI, then I also think that's unlikely by 2030.

One thing I should emphasise, and maybe we should have emphasised more, is the possibility of faster (say 10x) growth rates from continued AI progress. Where I think the best accounts of why this could happen do not depend on AI becoming superhuman in some very comprehensive way, or on a software-only intelligence explosion happening. The main thing we were trying to argue is that sustained AI progress could significantly accelerate tech progress and growth, after the point where (if it ever happens) the contribution to research from AI reaches some kind of parity with humans. I do think this is plausible, though not overwhelmingly likely.

Thanks again for writing this.

  1. ^

    That is, “AGI by 2030” or some other fixed calendar date.

  2. ^

    Things like memorisation, reading comprehension, abstract reasoning, general knowledge, etc.

Load more