This just came to mind: the reason that it's the wrong way to go about solving problems is that you want to solve the largest problems (well, per resource) and not just solve any random problem. Like, there is a problem that my shoes are currently untied, and I don't want to bend down or spend 10 seconds to tie them, but it's not very important.
So if you want to solve the most important problems, you should start with the problem and then work backwards for what solutions you might wish existed. I think the mere fact that people often talk about forecasting as the solution they are seeking to apply, whether that be Sentinel or whoever, is evidence that things are going wrong.
This just came to mind: the reason that it's the wrong way to go about solving problems is that you want to solve the largest problems (well, per resource) and not just solve any random problem. Like, there is a problem that my shoes are currently untied, and I don't want to bend down or spend 10 seconds to tie them, but it's not very important.
So if you want to solve the most important problems, you should start with the problem and then work backwards for what solutions you might wish existed. I think the mere fact that people often talk about forecasting as the solution they are seeking to apply, whether that be Sentinel or whoever, is evidence that things are going wrong.
Yea, this is fair. I am much more sympathetic to non-PM forecasting than I am PM/judgemental forecasting. The ideas in this post were really developed in 2023/2024 when I saw EAs spending a ton of time on Manifold/Metaculus, investing at high valuations, generally revering prediction markets for decision making, etc. whereas what I was seeing was completely different.
This post really belabours the first and second bullet point, perhaps because that is where a lot of money has gone to, but there can be a lot of value in the third.
I really believe in following the money. I think if we spend $100M on forecasting and $90M of it went to prediction market-style forecasting, I think it's fair to basically lump it all together. It'd be one thing if PMs were a small experiment within broad forecasting, but its been the main thing.
Hi Eva,
I think the Social Science Prediction Platform (alongside a friend of mine who is doing something similar for clinical trials) are among the more interesting uses of forecasting/PMs but I'm skeptical they will be uptaken to the degree/impact you might hope for.
do forecasts inform 1% of their funding or what?
I'm skeptical of things of the form "small percentage chance * big number". I think humans are really bad at estimating small percentages.
Would be happy to talk privately about any situations you are thinking of.
As promised, my reply (a couple days late).
I've seen a lot of posts that we need a lot more AI safety grantmakers. I feel like I want to do a bit of rough math and just see if that's the case. There is this estimate for the number of FTEs in AI safety by Stephen Mcaleese from Sept 2025 and 2022. Let's extrapolate exponential growth and say there are ~1400 FTEs on AI safety right now. Let's also assume from Julian Hazell's post that there are ~50 full-time AI safety grantmakers (though I think it's probably a bit more than that, given CG, Astralis, Astera, Longview, SFF, independent grantmakers, FLI, UK AISI, ARIA, AISTOF, Navigation Fundpeople at Schmidt, Macroscopic, etc., LTFF, Bluedot grants, Manifund, Tarbell, etc.).
From what I know about CG and other grantmakers, the people there are quite talented, and I would speculate are more talented than the average grantee.
Right off the bat, that means that right now, about 28 FTEs are working in AIS per grantmaker. Not to mention, a lot of the people who work full-time in AIS are working at frontier labs or other for-profit companies like Goodfire or in government (like UK AISI, CAISI), who don't need grantmakers to evaluate/fund their work. But we can ignore all those and just stick with the 28 FTE number.
I think I would expect the average grantmaker to be able to handle more than that, especially since an average organization usually has ~10 FTEs on average (I just asked Claude), and I expect a typical grantmaker to handle much more than 3 grants.
Also, I suspect a lot of grants look a lot more like renewals, and so don't need nearly as much review. For example, I'd expect grants to MATS and Redwood to look a lot more like reviewing their plans and signing off on them.
What am I missing?