Marcus Abramovitch 🔸

3913 karmaJoined

Comments
247

I've seen a lot of posts that we need a lot more AI safety grantmakers. I feel like I want to do a bit of rough math and just see if that's the case. There is this estimate for the number of FTEs in AI safety by Stephen Mcaleese from Sept 2025 and 2022. Let's extrapolate exponential growth and say there are ~1400 FTEs on AI safety right now. Let's also assume from Julian Hazell's post that there are ~50 full-time AI safety grantmakers (though I think it's probably a bit more than that, given CG, Astralis, Astera, Longview, SFF, independent grantmakers, FLI, UK AISI, ARIA, AISTOF, Navigation Fundpeople at Schmidt, Macroscopic, etc., LTFF, Bluedot grants, Manifund, Tarbell, etc.).

From what I know about CG and other grantmakers, the people there are quite talented, and I would speculate are more talented than the average grantee.

Right off the bat, that means that right now, about 28 FTEs are working in AIS per grantmaker. Not to mention, a lot of the people who work full-time in AIS are working at frontier labs or other for-profit companies like Goodfire or in government (like UK AISI, CAISI), who don't need grantmakers to evaluate/fund their work. But we can ignore all those and just stick with the 28 FTE number.

I think I would expect the average grantmaker to be able to handle more than that, especially since an average organization usually has ~10 FTEs on average (I just asked Claude), and I expect a typical grantmaker to handle much more than 3 grants.

Also, I suspect a lot of grants look a lot more like renewals, and so don't need nearly as much review. For example, I'd expect grants to MATS and Redwood to look a lot more like reviewing their plans and signing off on them.

What am I missing?

I just wanted to say, Phil's discussion with me in DMs has been very good about this and im going to be testing this too with some people.

I think he wrote this post off the cuff but this has been tremendously underdiscussed

This just came to mind: the reason that it's the wrong way to go about solving problems is that you want to solve the largest problems (well, per resource) and not just solve any random problem. Like, there is a problem that my shoes are currently untied, and I don't want to bend down or spend 10 seconds to tie them, but it's not very important.

So if you want to solve the most important problems, you should start with the problem and then work backwards for what solutions you might wish existed. I think the mere fact that people often talk about forecasting as the solution they are seeking to apply, whether that be Sentinel or whoever, is evidence that things are going wrong.

This just came to mind: the reason that it's the wrong way to go about solving problems is that you want to solve the largest problems (well, per resource) and not just solve any random problem. Like, there is a problem that my shoes are currently untied, and I don't want to bend down or spend 10 seconds to tie them, but it's not very important.

So if you want to solve the most important problems, you should start with the problem and then work backwards for what solutions you might wish existed. I think the mere fact that people often talk about forecasting as the solution they are seeking to apply, whether that be Sentinel or whoever, is evidence that things are going wrong.

One thing I'd flag is that models are extremely good at telling who is prompting them, and this leads to them being sycophantic, in very subtle ways. I'm not quite sure how they do it, but I've seen this in multiple instances.

For starters, you'd want to lobby for external safety testing on internal models.  You'd want to make sure external safety researchers had access to the models. You'd want certain reporting, etc.

I think he would include a lot of people who work at Anthropic, for example, on pre-training, some of whom went through MATS or something.

Yea, this is fair. I am much more sympathetic to non-PM forecasting than I am PM/judgemental forecasting. The ideas in this post were really developed in 2023/2024 when I saw EAs spending a ton of time on Manifold/Metaculus, investing at high valuations, generally revering prediction markets for decision making, etc. whereas what I was seeing was completely different.

This post really belabours the first and second bullet point, perhaps because that is where a lot of money has gone to, but there can be a lot of value in the third.

I really believe in following the money. I think if we spend $100M on forecasting and $90M of it went to prediction market-style forecasting, I think it's fair to basically lump it all together. It'd be one thing if PMs were a small experiment within broad forecasting, but its been the main thing.

Hi Eva,

I think the Social Science Prediction Platform (alongside a friend of mine who is doing something similar for clinical trials) are among the more interesting uses of forecasting/PMs but I'm skeptical they will be uptaken to the degree/impact you might hope for.

do forecasts inform 1% of their funding or what?

I'm skeptical of things of the form "small percentage chance * big number". I think humans are really bad at estimating small percentages.

Would be happy to talk privately about any situations you are thinking of.

As promised, my reply (a couple days late).

  1. I think it's a bit of a cop out to suggest that the money was spent poorly and therefore it's not fair to judge forecasting on the merits. Not sure if this is what you are saying though.
  2. I'm not sure I agree. Blockchains were funded. Lots of academic research is funded by governments, etc.
  3. By this logic, why aren't we paying people in the top, say 100 on Polymarket on non-sports questions and saying "hey guys, we want to pay you to make some forecasts for us".
  4. I'm not sure how I feel about these. I will say, I am not nearly as big a fan of AI 2027 as others seem to be, and I think it is going to severely discredit AI risks because we have been crying wolf when frankly, most of the scary stuff they say doesn't happen. I am very happy to make some bets with Daniel or Eli on some of these (and give them extra time).
  5. Useless is a stretch and I didn't intend to claim it originally. Sorry if I implied it. I merely think they are overrated, hence the title.
Load more