Take the 2025 EA Forum Survey to help inform our strategy and prioritiesTake the survey
Hide table of contents

Edit 11 Feb 2022: Jeremy made a post about starting a low-commitment LW Article club where he'd be linkposting articles on a weekly basis from this list for people to engage with! 

Context / Motivation: 

  • I am interested in thinking a lot recently about how we could share ideas from the rationalist community to EA and related subcommunities (perhaps communicating the same ideas in different ways).
  • I've been diving into LessWrong recently and remembered why I hadn't for a while - it's really overwhelming. Even with the sequences and curation, it's a lot of content, and it's not always obvious to me which posts I'd find most valuable.
  • I think it's better to read fewer posts in more depth to properly understand them.
  • I think it's likely that some posts or ideas will be much more relevant to EAs than others, but I'm not sure which ones

My ask:

  • I'd be interested in recommendations for standlone posts (e.g. All debates are bravery debates), specific concepts (e.g. schelling fence or doublecrux) , or specific sequences (e.g.)
  • If you have time, I'd love to know why it's valuable to you
New Answer
New Comment


8 Answers sorted by

I'm a big fan of some of the early LessWrong content, e.g.

More generally, I'd recommend much of the content by Scott Alexander ("Yvain"), Paul Christiano, Wei Dai, Gwern, Greg Lewis ("Thrasymachus"), Anna Salamon and Carl Shulman (I'm probably forgetting other names).

Privileging the Question changed my life in college. I don't know how useful it would be for the average person already involved in EA, but it played a huge role in my not getting distracted by random issues and controversies, and instead focusing on big-picture problems that weren't as inherently interesting. I'd at least recommend it to new members of university EA groups, if not "most community members".

This got me to leave my girlfriend and has remained a permanent way that I think:

https://www.lesswrong.com/posts/627DZcvme7nLDrbZu/update-yourself-incrementally

 

I read it as part of all the sequences, no idea how helpful it will be to others or as a standalone post

My take is that LessWrong is best understood as a mix of individual voices, each with their own style and concerns. The approach I'd recommend is to select one writer whose voice you find compelling, and spend some time digesting their ideas. A common refrain is "read the sequences," but that's not where I started. I like John Wentworth's writing.

Alternatively, you might find yourself interested in a particular topic. LessWrong's tags can help you both find an interesting topic and locate relevant posts, though it's not super fine grained or comprehensive.

One of the key sources of value on LessWrong is that it provides a common language for some complex ideas, presented in a relatively fun and accessible format. The combination of all those ideas can elevate thinking, although it's no panacaea. My intuition is that it's best to slowly follow your curiosity over a period of a few years, rather than trying to digest the whole thing all at once, or pick a couple highlights.

Any particular Wentworth posts that stand out to you? I'd like to include some in the LCLWBC (full credit to you for the name!), but I am not too familiar.

John had several posts highly ranked in the 2020 LessWrong review, and one in the 2019 LessWrong review, so there's a community consensus that they're good. There was also a 2018 LessWrong review, though John didn't place there.

In general, the review is a great resource for navigating more recent LW content. Although old posts are a community touchstone, the review includes posts that reflect the live interests of community members that have also been extensively vetted not only for being exciting, but for maintaining their value a year later.

Thank you!

I really like Ends Don't Justify Means (Among Humans) and think it's a bit underrated. (In that I don't hear people reference it much.)

I think I find the lesson generally useful: that in some cases it can be bad for me to "follow consequentialism," (because in some cases I'm an idiot) without consequentialism being itself bad.

The noncentral fallacy nicely categorizes a very common source of ethical disagreement in my experience.

[Edit:] Somewhat more niche, but considering how important AI risk is to many EAs, I'd also recommend Against GDP as a metric for timelines and takeoff speeds, for rebutting what is in my estimation a bizarrely common error in forecasting AI takeoff.

Comments9
Sorted by Click to highlight new comments since:

Maybe the thing to do would be to start a low-commitment LW book club? There's so many old posts that it doesn't feel fresh to comment on them, but having a way to put some group attention on a couple posts at a time might help.

I made a separate post to get the ball rolling and make sure this happens. 

Would love to do this !

Agreed - would love to participate in something like this, and would encourage other group members (esp. organizers) to as well!

I'd be interested in something like this. 

I'd also be interested in pursuing this idea! LW can definitely be overwhelming, and it'd be a fun (and useful) project to take a deep dive and perhaps produce a recommended reading list for others (broadly defined).

It took me a while to get rolling, but I have done a first Less Wrong repost here and will continue weekly as long as there is enough interest. 

Saved these all to pocket, thanks for the recommendations! 

Curated and popular this week
 ·  · 1m read
 · 
This morning I was looking into Switzerland's new animal welfare labelling law. I was going through the list of abuses that are now required to be documented on labels, and one of them made me do a double-take: "Frogs: Leg removal without anaesthesia."  This confused me. Why are we talking about anaesthesia? Shouldn't the frogs be dead before having their legs removed? It turns out the answer is no; standard industry practice is to cut their legs off while they are fully conscious. They remain alive and responsive for up to 15 minutes afterward. As far as I can tell, there are zero welfare regulations in any major producing country. The scientific evidence for frog sentience is robust - they have nociceptors, opioid receptors, demonstrate pain avoidance learning, and show cognitive abilities including spatial mapping and rule-based learning.  It's hard to find data on the scale of this issue, but estimates put the order of magnitude at billions of frogs annually. I could not find any organisations working directly on frog welfare interventions.  Here are the organizations I found that come closest: * Animal Welfare Institute has documented the issue and published reports, but their focus appears more on the ecological impact and population decline rather than welfare reforms * PETA has conducted investigations and released footage, but their approach is typically to advocate for complete elimination of the practice rather than welfare improvements * Pro Wildlife, Defenders of Wildlife focus on conservation and sustainability rather than welfare standards This issue seems tractable. There is scientific research on humane euthanasia methods for amphibians, but this research is primarily for laboratory settings rather than commercial operations. The EU imports the majority of traded frog legs through just a few countries such as Indonesia and Vietnam, creating clear policy leverage points. A major retailer (Carrefour) just stopped selling frog legs after welfar
 ·  · 4m read
 · 
Note: This post was crossposted from the Open Philanthropy Farm Animal Welfare Research Newsletter by the Forum team, with the author's permission. The author may not see or respond to comments on this post. ---------------------------------------- > Why ending the worst abuses of factory farming is an issue ripe for moral reform I recently joined Dwarkesh Patel’s podcast to discuss factory farming. I hope you’ll give it a listen — and consider supporting his fundraiser for FarmKind’s Impact Fund. (Dwarkesh is matching all donations up to $250K; use the code “dwarkesh”.) We discuss two contradictory views about factory farming that produce the same conclusion: that its end is either inevitable or impossible. Some techno-optimists assume factory farming will vanish in the wake of AGI. Some pessimists see reforming it as a hopeless cause. Both camps arrive at the same conclusion: fatalism. If factory farming is destined to end, or persist, then what’s the point in fighting it? I think both views are wrong. In fact, I think factory farming sits in the ideal position for moral reform. Because its end is neither inevitable nor impossible, it offers a unique opportunity for advocacy to change the trajectory of human moral progress. Not inevitable Dwarkesh raised an objection to working on factory farming that I often hear from techno-optimists who care about the issue: isn’t its end inevitable? Some cite the long arc of moral progress; others the promise of vast technological change like cultivated meat or Artificial General Intelligence (AGI) which surpasses human capabilities. It’s true that humanity has achieved incredible moral progress for humans. But that progress was never inevitable — it was the result of moral and political reform as much as technology. And that moral progress mostly hasn’t yet extended to animals. For them, the long moral arc of history has so far only bent downward. Technology may one day end factory farming, just as cars liberated w
 ·  · 1m read
 · 
This is a personal essay about my failed attempt to convince effective altruists to become socialists. I started as a convinced socialist who thought EA ignored the 'root causes' of poverty by focusing on charity instead of structural change. After studying sociology and economics to build a rigorous case for socialism, the project completely backfired as I realized my political beliefs were largely psychological coping mechanisms. Here are the key points: * Understanding the "root cause" of a problem doesn't necessarily lead to better solutions - Even if capitalism causes poverty, understanding "dynamics of capitalism" won't necessarily help you solve it * Abstract sociological theories are mostly obscurantist bullshit - Academic sociology suffers from either unrealistic mathematical models or vague, unfalsifiable claims that don't help you understand or change the world * The world is better understood as misaligned incentives rather than coordinated oppression - Most social problems stem from coordination failures and competing interests, not a capitalist class conspiring against everyone else * Individual variation undermines class-based politics - People within the same "class" have wildly different cognitive traits, interests, and beliefs, making collective action nearly impossible * Political beliefs serve important psychological functions - They help us cope with personal limitations and maintain self-esteem, often at the expense of accuracy * Evolution shaped us for competition, not truth - Our brains prioritize survival, status, and reproduction over understanding reality or being happy * Marx's insights, properly applied, undermine the Marxist political project - His theory of ideological formation aligns with evolutionary psychology, but when applied to individuals rather than classes, it explains why the working class will not overthrow capitalism. In terms of ideas, I don’t think there’s anything too groundbreaking in this essay. A lot of the