Hide table of contents

This is a retrospective post for the Tools for Collaborative Truth Seeking project. I recommend reading it if you want to do a project involving writing tutorials for the forum, matched forum posts and live streams, or short projects involving user interviews. 

Some quick notes: the project took around 55 hours of work in total, and consisted of 8 forum posts and 8 live events in the EA GAtherTown where we discussed the use of the tool and used it for a short exercise. The posts were put up once a day over a week and a half, with the live streams initially the same day as the post, but by halfway we changed to the day after. 

I feel the two key mistakes were not understanding where the list of tools had come from and not re-assessing as the project grew. I'll also mention a couple of more specific points about the format and timing of the final product. 

Key Mistakes

I feel the biggest mistake I made with this project was not asking where the list of tools had come from. Even assuming it was from a great source, not being able to regenerate the method really limited my ability to write about the true reasons why a tool was on the list, and how it compared to other tools[1]

The other really central mistake was that the project changed quite a lot over time, and we never stopped to re-assess. Initially, the project was just the wiki page, then we added a sequence of posts to explain the tools, then we added livestreams.  At some point, the project had changed enough that we should have stopped and looked at the project with fresh eyes. 

I think this affected the quality of the final product; it was less cohesive than it should have been, and the timing was a little off[2]. An (untested) rule of thumb I will be trying to implement is to re-assess all of your assumptions about a project when you make a decision that seems to at least double the time the project will take.

More Specific Mistakes

Scheduling

The tight schedule felt good generally to keep people interested, but the biggest scheduling mistake was that people didn't have enough time between the post going up and the event itself. Our best-attended event was Squiggle, the post for which went up a few days in advance of the event due to the weekend (though this was also our most popular post). I suggest posting event announcements right at the beginning of the project, so people can book time into their day to come to live events.

The other minor issue was that some people who came to several events said they wished they had been spaced out slightly more; maybe every other day. I think this would be particularly important for projects longer than ours or where you're expecting the same people to turn up to several live events. 

Livestreams

Attendance for some livestreams was low (0-1). The format of having an exercise to do seemed useful. The main mistake was not bringing the devs of the tools in earlier and advertising their attendance and/or getting them more involved in picking exercises that showcased the tools well.

User Interviews

User interviews were run less like writing feedback and more like software user testing. This worked well when we were able to get access to people who were the target audience, and was not so helpful otherwise; I recommend putting effort into reaching your target audience for this kind of interview. People who understood this was the type of feedback we were looking for were also more helpful.

LessWrong

The posts generally did very poorly on LessWrong. This could be because the titles followed more of the forum format (very explicit, not weird at all), or because LW is more about working on AI than the forum. I think the two mistakes here were not talking to the LessWrong team about the project and just cross-posting rather than having a version tailored to LessWrong (e.g. with more AI examples, more individual working).

Conclusion

Overall, I was quite happy with how the project went. The key metrics will be how many people are using these tools in 6 months and what they're using them for, but I'm cautiously optimistic. The key mistakes were not re-assessing the project when it grew, and not thinking carefully about how we formed the list of tools that were included

  1. ^

    For those interested, the initial list came from a brainstorm on an epistemics slack, though it was altered quite substantially through the course of the project

  2. ^

    e.g. I wish we had posted all of the event announcements a week or two in advance of the posts going up, so people could plan ahead

Comments4


Sorted by Click to highlight new comments since:

Thanks for doing this! I appreciate the willingness to think about mistakes, but for what it's worth I would also be interested to hear what things well. At the end you allude to "key metrics" that haven't resolved yet, but it might be worth sharing what the initial results are?

Hi! Sorry for the delay in replying -- we've now posted the metrics, if you're interested.
 

I just want to shout out and say that, although my only interaction with the series has been through reading the posts, it has had a big impact for me and my workplace. Several coworkers and I now regularly use Excalidraw and Guesstimate. This has shaved about ~1 week off of a large project. Just napkin math here but we estimate our project will save 50 lives/week with a 1 in 10 chance of success. So by those numbers I'd say you saved about 5 lives with this project. Anyway, just napkin math, but I just wanted to say thank you!

Thanks for sharing!

Curated and popular this week
 ·  · 25m read
 · 
Epistemic status: This post — the result of a loosely timeboxed ~2-day sprint[1] — is more like “research notes with rough takes” than “report with solid answers.” You should interpret the things we say as best guesses, and not give them much more weight than that. Summary There’s been some discussion of what “transformative AI may arrive soon” might mean for animal advocates. After a very shallow review, we’ve tentatively concluded that radical changes to the animal welfare (AW) field are not yet warranted. In particular: * Some ideas in this space seem fairly promising, but in the “maybe a researcher should look into this” stage, rather than “shovel-ready” * We’re skeptical of the case for most speculative “TAI<>AW” projects * We think the most common version of this argument underrates how radically weird post-“transformative”-AI worlds would be, and how much this harms our ability to predict the longer-run effects of interventions available to us today. Without specific reasons to believe that an intervention is especially robust,[2] we think it’s best to discount its expected value to ~zero. Here’s a brief overview of our (tentative!) actionable takes on this question[3]: ✅ Some things we recommend❌ Some things we don’t recommend * Dedicating some amount of (ongoing) attention to the possibility of “AW lock ins”[4]  * Pursuing other exploratory research on what transformative AI might mean for animals & how to help (we’re unconvinced by most existing proposals, but many of these ideas have received <1 month of research effort from everyone in the space combined — it would be unsurprising if even just a few months of effort turned up better ideas) * Investing in highly “flexible” capacity for advancing animal interests in AI-transformed worlds * Trying to use AI for near-term animal welfare work, and fundraising from donors who have invested in AI * Heavily discounting “normal” interventions that take 10+ years to help animals * “Rowing” on na
 ·  · 3m read
 · 
About the program Hi! We’re Chana and Aric, from the new 80,000 Hours video program. For over a decade, 80,000 Hours has been talking about the world’s most pressing problems in newsletters, articles and many extremely lengthy podcasts. But today’s world calls for video, so we’ve started a video program[1], and we’re so excited to tell you about it! 80,000 Hours is launching AI in Context, a new YouTube channel hosted by Aric Floyd. Together with associated Instagram and TikTok accounts, the channel will aim to inform, entertain, and energize with a mix of long and shortform videos about the risks of transformative AI, and what people can do about them. [Chana has also been experimenting with making shortform videos, which you can check out here; we’re still deciding on what form her content creation will take] We hope to bring our own personalities and perspectives on these issues, alongside humor, earnestness, and nuance. We want to help people make sense of the world we're in and think about what role they might play in the upcoming years of potentially rapid change. Our first long-form video For our first long-form video, we decided to explore AI Futures Project’s AI 2027 scenario (which has been widely discussed on the Forum). It combines quantitative forecasting and storytelling to depict a possible future that might include human extinction, or in a better outcome, “merely” an unprecedented concentration of power. Why? We wanted to start our new channel with a compelling story that viewers can sink their teeth into, and that a wide audience would have reason to watch, even if they don’t yet know who we are or trust our viewpoints yet. (We think a video about “Why AI might pose an existential risk”, for example, might depend more on pre-existing trust to succeed.) We also saw this as an opportunity to tell the world about the ideas and people that have for years been anticipating the progress and dangers of AI (that’s many of you!), and invite the br
 ·  · 14m read
 · 
Cross-posted from my personal blog after strong encouragement from two people. The text is partially personal musings over my journey through the pre-surgery process, and partially a review of arguments why it might or might not make sense for a person who cares about effective altruism to donate a kidney. Some of the content is specific to where I am based in (Finland), and some of the content is less honed than I'd like, but if I didn't timebox my writing and just push whatever I have at the end of it I probably would not publish anything. Hope you enjoy! June 4, 2025 I was eating phở in the city center when the call came. The EM study had been done. "Bad news," my nephrologist said. "You have a kidney disease." The words hit harder than the spicy broth. After more than nine months of tests, blood draws, and even a kidney biopsy, my journey to donate a kidney to my friend had just come to an abrupt end. Let’s rewind to where this all began. The Decision When Effective Altruism volunteer Mikko mentioned that his kidney transplant was showing signs of chronic rejection, I jumped and offered him my kidney. I had read Dylan Matthew’s kidney donation story, Scott Alexander’s kidney donation story, and had taken part in a discussion on estimating the level of kidney demand in Finland in a local EA discussion group that had at least one Finnish doctor involved. The statistics were reassuring. Kidney donations are very safe with a perioperative death rate of around 0.03% to 0.06%. You will have a slightly increased risk of kidney-related issues later in life (1-2% likelihood of kidney failure), and there is also some weak evidence that suggests people who donate a kidney have very slightly shorter life expectancy than those who have both kidneys intact. The QALY Trade-off There are multiple attempts to calculate what the quality-adjusted life year (QALY) trade for a kidney donation actually is, and to me it looks like donating a kidney is by default similar to