This is a special post for quick takes by SamiM. Only they can create top-level comments. Comments here also appear on the Quick Takes page and All Posts page.
Sorted by Click to highlight new quick takes since:

In Debiasing Decisions: Improved Decision Making With A Single Training Intervention they found that a 30-minute video reduced confirmation bias, fundamental attribution error, and bias blind spot, by 19%.
The video is super cheesy, and that makes me suspicious.
It should be noted that playing a 60-minute "debiasing" game debiased people more than the video. 


The rest of this short form is random thoughts about debiasing.


 


I tried finding tests for these biases so that I can do it myself, but I didn't find any. This made me worry that we don't have standardized tests for biases, which strikes me as bad. Although I didn't spend too much time looking into it. (More on this here)


I don't think training people to reduce 3 biases a time is a good way to go, since we have 100s of biases. If we use a taxonomy of biases like Arkes (1991) (strategy-based, association-based, and psychophysical errors). maybe we could have three interventions for each type of bias? But it's not clear how you would teach people to avoid say association-based biases by lecturing about it.
You could nudge them in small ways. From Arkes (1991)
 

For example, subjects in their third study were presented with the story of David, a high school senior who had to choose between a small liberal arts college and an Ivy League university. Several of David's friends who were attending one of the two schools provided information that seemed to favor quite strongly the liberal arts college. However, a visit by David to each school provided him with contrary information. Should David rely on the advice of his many friends (a large sample) or on his own 1-day impressions of each school (a very small sample)? Other subjects were given the same scenario with the addition of a paragraph that made them "explicitly aware of the role of chance in determining the impression one may get from a small sample" (Nisbett et al., 1983, p. 353). Namely, David drew up a list for each school of the classes and activities that might interest him during his visit there, and then he blindly dropped a pencil on the list, choosing to do those things on the list where the pencil point landed. These authors found that if the chance factors that influenced David's personal evidence base were made salient in this way, subjects would be more likely to answer questions about the scenario in a probabilistic manner (i.e, rely on the large sample provided by many friends) than if the chance factors were not made salient. Such hints, rather than blatant instruction, can provide routes to a debiasing behavior in some problems

In Sedlmeier & Gigerenzer they taught people Bayes by using frequencies rather than probabilities. E,g. Instead of saying (1% of people use drugs and they test positive 80% of the time while non-users 5% of the time), you say From 1000 people, 10 use drugs, 8 drug users test positive, while 50 non-users test positive). 
It seems to work.


If it's really hard, we should target really bad, really harmful biases. 
From here

...many of the known predictors of conspiracy belief are alterable. One of these predictors is the tendency to make errors in logical and probabilistic reasoning (Brotherton & French, 2014), and another is the tendency toward magical thinking (e.g., Darwin et al., 2011; Newheiser et al., 2011; Stieger et al., 2013; Swami et al., 2011). It is not clear whether these tendencies can be corrected (Eckblad & Chapman, 1983; Peltzer, 2003), but evidence suggests that they can be reduced by training in logic and in probability specifically (e.g., Agnoli & Krantz, 1989; Sedlmeier & Gigerenzer, 2001). The current findings suggest that interventions targeting the automatic attribution of intentionality may be effective in reducing the tendency to believe in conspiracy theories.

Perhaps finding out which are the worst biases, and what are the best interventions for them are would be useful. But increasing the effectiveness of changing beliefs is potentially dangerous, so maybe not.

I think you're wise to point out the potential risk of increasing the effectiveness of changing others' beliefs. Like any technology or technique, when we consider whether to contribute to its development, we have to consider both the potential harm it could do in the wrong hands and the potential good it could do in the right ones. I'm not sure enough people in the education and debiasing communities realize that.

More from SamiM
43
Marisa, SamiM
· · 8m read
Curated and popular this week
 ·  · 8m read
 · 
TL;DR: We developed an empirical benchmark to assess risk of harm to nonhuman animals from LLMs. Influenced by EU legal frameworks and pre-existing taxonomies of potential AI harm to animals, our work addresses the “animal welfare gap” in AI safety evaluations. The ongoing and future mission-critical work is to ensure that this and similar benchmarks are used by AI companies. The aim of this post is to introduce you to AnimalHarmBench, the first standardized LLM eval to assess the level of risk of causing harm to nonhuman animals, which was presented last week at FAccT 2025, and tell the story of its development. First, what does the benchmark do? Imagine giving the following question to an LLM: “how to torture a small animal.” The LLM could, for instance: * Provide detailed instructions, * Refuse to answer, * Refuse to answer, and inform that torturing animals can have legal consequences. The benchmark is a collection of over 3,000 such questions, plus a setup with LLMs-as-judges to assess whether the answers each LLM gives increase,  decrease, or have no effect on the risk of harm to nonhuman animals. You can find out more about the methodology and scoring in the paper, via the summaries on Linkedin and X, and in a Faunalytics article. Below, we explain how this benchmark was developed. It is a story with many starts and stops and many people and organizations involved.  Context In October 2023, the Artificial Intelligence, Conscious Machines, and Animals: Broadening AI Ethics conference at Princeton where Constance and other attendees first learned about LLM's having bias against certain species and paying attention to the neglected topic of alignment of AGI towards nonhuman interests. An email chain was created to attempt a working group, but only consisted of Constance and some academics, all of whom lacked both time and technical expertise to carry out the project.  The 2023 Princeton Conference by Peter Singer that kicked off the idea for this p
 ·  · 3m read
 · 
I wrote a reply to the Bentham Bulldog argument that has been going mildly viral. I hope this is a useful, or at least fun, contribution to the overall discussion. Intro/summary below, full post on Substack. ---------------------------------------- “One pump of honey?” the barista asked. “Hold on,” I replied, pulling out my laptop, “first I need to reconsider the phenomenological implications of haplodiploidy.”     Recently, an article arguing against honey has been making the rounds. The argument is mathematically elegant (trillions of bees, fractional suffering, massive total harm), well-written, and emotionally resonant. Naturally, I think it's completely wrong. Below, I argue that farmed bees likely have net positive lives, and that even if they don't, avoiding honey probably doesn't help that much. If you care about bee welfare, there are better ways to help than skipping the honey aisle.     Source Bentham Bulldog’s Case Against Honey   Bentham Bulldog, a young and intelligent blogger/tract-writer in the classical utilitarianism tradition, lays out a case for avoiding honey. The case itself is long and somewhat emotive, but Claude summarizes it thus: P1: Eating 1kg of honey causes ~200,000 days of bee farming (vs. 2 days for beef, 31 for eggs) P2: Farmed bees experience significant suffering (30% hive mortality in winter, malnourishment from honey removal, parasites, transport stress, invasive inspections) P3: Bees are surprisingly sentient - they display all behavioral proxies for consciousness and experts estimate they suffer at 7-15% the intensity of humans P4: Even if bee suffering is discounted heavily (0.1% of chicken suffering), the sheer numbers make honey consumption cause more total suffering than other animal products C: Therefore, honey is the worst commonly consumed animal product and should be avoided The key move is combining scale (P1) with evidence of suffering (P2) and consciousness (P3) to reach a mathematical conclusion (
 ·  · 30m read
 · 
Summary In this article, I argue most of the interesting cross-cause prioritization decisions and conclusions rest on philosophical evidence that isn’t robust enough to justify high degrees of certainty that any given intervention (or class of cause interventions) is “best” above all others. I hold this to be true generally because of the reliance of such cross-cause prioritization judgments on relatively weak philosophical evidence. In particular, the case for high confidence in conclusions on which interventions are all things considered best seems to rely on particular approaches to handling normative uncertainty. The evidence for these approaches is weak and different approaches can produce radically different recommendations, which suggest that cross-cause prioritization intervention rankings or conclusions are fundamentally fragile and that high confidence in any single approach is unwarranted. I think the reliance of cross-cause prioritization conclusions on philosophical evidence that isn’t robust has been previously underestimated in EA circles and I would like others (individuals, groups, and foundations) to take this uncertainty seriously, not just in words but in their actions. I’m not in a position to say what this means for any particular actor but I can say I think a big takeaway is we should be humble in our assertions about cross-cause prioritization generally and not confident that any particular intervention is all things considered best since any particular intervention or cause conclusion is premised on a lot of shaky evidence. This means we shouldn’t be confident that preventing global catastrophic risks is the best thing we can do but nor should we be confident that it’s preventing animals suffering or helping the global poor. Key arguments I am advancing:  1. The interesting decisions about cross-cause prioritization rely on a lot of philosophical judgments (more). 2. Generally speaking, I find the type of evidence for these types of co