AI safety researcher
If your algorithms get more efficient over time at both small and large scales, and experiments test incremental improvements to architecture or data, then they should get cheaper to run proportionally to algorithmic efficiency of cognitive labor. I think this is better as a first approximation than assuming they're constant, and might hold in practice especially when you can target small-scale algorithmic improvements.
I'm worried that trying to estimate by looking at wages is subject to lots of noise due to assumptions being violated, which could result in the large discrepancy you see between the two estimates.
One worry: I would guess that Anthropic could derive more output from extra researchers (1.5x/doubling?) than from extra GPUs (1.18x/doubling?), yet it spends more on compute than researchers. In particular I'd guess alpha/beta = 2.5, and wages/r_{research} is around 0.28 (maybe you have better data here). Under Cobb-Douglas and perfect competition these should be equal, but they're off by a factor of 9! I'm not totally sure but I think this would give you strange parameter values in CES as well. This huge gap between output elasticity and where firms are spending their money is strange to me, so I strongly suspect that one of the assumptions is broken rather than just being some extreme value like -0.10 or 2.58 with large firm fixed effects.
My guess at why: The AI industry is very different than it was in 2012 so it is plausible these firm fixed effects have actually greatly changed over time, which would affect the regression coefficients. Just some examples of possible changes over time:
Nevertheless I'm excited about the prospect of estimating and and I'm glad this was posted. Are you planning follow-up work, or is there other economic data we could theoretically collect that could give us higher confidence estimates?
(edited to fix numbers, I forgot 2 boxes means +3dB)
dB is logarithmic so a proportional reduction in sound energy will mean subtracting an absolute number of dB, not a percentage reduction in dB.
HouseFresh tested the AirFanta 3Pro https://housefresh.com/airfanta-3pro-review/ at different voltage levels and found:
So basically you subtract 13 dB when halving the CADR. I now realize that if you have two boxes, the sound energy will double (+3dB) and so you'll actually only get -10 dB from running two at half speed. So a more accurate statement for the Airfanta would be that for -15dB noise at the same CADR, you need something like 2.8 purifiers running at 36% speed. It's still definitely possible to markedly lower noise by adding more filter area.
Your box fan CR box data tell a similar story. If logarithmic scaling is accurate, the sound reduction for halving CADR would be ln(1/2)/ln(165/239)*(8 dB) = 15 dB, or 12 dB for maintaining CADR with double the units. It just doesn't have a speed low enough to get these low noise levels (and due to the box fan's low static pressure you might need to add more filters per fan at low speeds).
Airfanta's absolute noise levels are high for a CR box type design but this is a device that retails for 298 CNY = $41 USD in China, runs at high speed, and uses near-HEPA (95%) rather than MERV filters so is to be expected.
I broadly agree with section 1, and in fact since we published I've been looking into how time horizon varies between domains. Not only is there lots of variance in time horizon, the rate of increase also varies significantly.
See a preliminary graph plus further observations on LessWrong shortform.
That's a box fan CR box; the better design (and the one linked) uses PC fans which are better optimized for noise. I don't have much first-hand experience with this, but physics suggests that noise from the fan will be proportional to power usage, which is pressure * airflow, if efficiency is constant, and this is roughly consistent with various tests I've found online.
Both further upsizing and better sound isolation would be great. What's the best way to reduce duct noise in practice? Is an 8" flexible duct quieter than a 6" rigid duct or will most of the noise improvement come from oversizing the termination, removing tight bends or installing some kind of silencer device? I might suggest this to a relative.
Isn't particulate what we care about? The purpose of the filters is to get particulate out of the air, and the controlled experiment Jeff did basically measures that. If air mixing is the concern, ceiling fans can mix air far more than required, and you can just measure particulate in several locations anyway.
One problem is putting everything on a common scale when historical improvements are so sensitive to the distribution of tasks. A human with a computer with C, compared to a human with just log tables, is a billion times faster at multiplying numbers but less than twice as fast at writing a novel. So your distribution of tasks has to be broad enough that it captures the capabilities you care about, but it also must be possible to measure a baseline score at low tech level and have a wide range of possible scores. This would make the benchmark extremely difficult to construct in practice.