Unable to work. Was community director of EA Netherlands, had to quit due to ME/CFS (presumably long covid). Everything written since 2021 with considerable brain fog, and bad at maintaining discussions/replying to comments since.
I have a background in philosophy, risk analysis, and moral psychology. I also did some x-risk research. Currently most worried about AI and US democracy. Interested in biomedical R&D reform.
It seems like a worthwhile project to ask/pressure Anthropic's founders to make their pledges legally binding.
Anthropic's founders have pledged to donate 80% of their wealth. Ozzie Gooen estimates that in a few years this could be worth >$40 billion.
As Ozzie writes, adherence to the Giving Pledge (the Gates one) is pretty low: only 36% of deceased original pledgers met the 50% commitment. It's hard to follow through on such commitments, even for (originally) highly morally motivated people.
This study suggests that, if done well, LLMs can estimate effect sizes of drugs fairly accurately based on online reports.
Sounds like this would benefit from this method of analyzing online anecdotes to estimate effect sizes, which showed good results in early data and is being expanded:
I sometimes think of this idea and haven't found anyone mentioning it with a quick AI search: a tax on suffering.
EDIT: there's a paper on this but specific to animal welfare that was shared on the forum earlier this year.
A suffering tax would function as a Pigouvian tax on negative externalities—specifically, the suffering imposed on sentient beings. The core logic: activities that cause suffering create costs not borne by the actor, so taxation internalizes these costs and incentivizes reduction.
This differs from existing approaches (animal welfare regulations, meat taxes) by:
The main problems are measurement & administration. I would imagine an institute would be tasked with guidelines/a calculation model, which could become pretty complex. Actually administrating it would also be very hard, and there should be a threshold beneath which no tax is required because it wouldn't be worth the overhead. I would imagine that an initial version wouldn't right away be "full EA" taking into account invertebrates. It should start with a narrow scope, but with the infrastructure for moral circle expansion.
It's obviously more a theoretical exercise than practical near-term, but here's a couple of considerations:
Couple of thoughts, kinda long and rambly
This gets discussed occasionally on the Manifold Discord and I wanted to share some skeptical points that one of the top forecasters (Semiotic Rivalry) made there:
This was largely in response to me saying that I find it hard to think through Trump/MAGA military (self-)coup possibilities. Because although military self-coups appear to be rare in consolidated or backsliding democracies, they're not entirely unheard of, and it sure seems like Hegseth, Trump, etc. are working towards this. They are systematically dismantling military guardrails:
The general pattern of purging appears to be that Trump/the administration gives an illegal/norm-breaking order which functions as a loyalty test: it forces everyone involved to either comply, step down, or refuse to obey (which tends to get you fired - something that the Supreme Court hasn't been adequately protecting besides the Fed).
The coup form I expect, if it happens, would not be a direct command to military generals, but to order his most loyal militarized groups (e.g. red states National Guard, ICE) to take control of the democratic/election process. Opposing military would then have to coordinate on action, which would be very difficult. The general population could resist en masse (South Korea 2024-style), but so far protests have been small, and in the US there's a vocal and dangerous base supporting Trump. That said, base rates suggest a coup is still very unlikely, and coups are difficult. I don't know what probability I would give it, I'm mainly trying to understand the mechanisms here.
Other thoughts:
Pet peeve: stop calling short timelines "optimistic" and long timelines "pessimistic". These create unwarranted connotations that day AI progress is desirable. Most people concerned about AI safety find short timelines dangerous! Instead, use "bullish" vs. "bearish", or just "short timelines" vs. "long timelines".
Ah, good post! I should probably just have refrained from commenting on the design given my limited knowledge. For benchmarking purposes, perhaps these designs are better than the one from Wirecutter:
Thanks, fixed the link