One of my goals for this year was to write a piece setting out my own views on the topic of AI consciousness and the possible paths forward. I wrote it because I realized that I'd co-authored a string of pieces (Butlin et al. 2023; Long et al. 2024; Keeling et al. 2024; Caviola et al. 2025; Colombatto et al. 2025) and yet none of these quite captured my personal position as of 2025. I've also given media interviews but they're always a bad way to explain your views clearly. 

To make matters worse, I've seen over the past year extreme positions on both sides becoming entrenched. The debate has been acquiring some of the features of a toxic debate - two sides aggressively mocking each other, taking offence at each other's choice of words - when I think the best approach goes calmly down the middle, acknowledging reasonable points both sides.

My preprint "AI Consciousness: A Centrist Manifesto" might be seen as an attempt to pull the debate back from the brink and return it to being a constructive, collaborative, mutually supportive endeavour. We need two research programmes in parallel: one aimed at helping ordinary users avoid unwarranted inferences to the human-like consciousness of their AI friends, assistants and partners, and another aimed at developing better tests for genuine forms of consciousness in AI. These forms of consciousness, if they are there at all, will be of a profoundly alien, radically un-human-like kind. These two research programs don't need to come into conflict. They can be pursued in mutually consistent ways.

The road to better tests for consciousness than we have now is long and daunting - as shown in the figure below - but we should reject any inference from long and daunting to "impossible". 

Here is a link to the paper: https://philpapers.org/rec/BIRACA-4

Here is the abstract: We face two urgent challenges concerning consciousness and AI. Challenge One is that millions of users will soon misattribute human-like consciousness to AI friends, partners, and assistants on the basis of mimicry and role-play, and we don’t know how to prevent this. Challenge Two is that profoundly alien forms of consciousness might genuinely be achieved in AI, but our theoretical understanding of consciousness is too immature to provide confident answers one way or the other. Centrism about AI consciousness is the position that we must take both challenges seriously. The two challenges interact in ways that make this difficult. Steps to address Challenge One might undermine attempts to address Challenge Two by portraying the idea of conscious AI as impossible or inherently unlikely. Conversely, attempts to address Challenge Two might lead to higher levels of misattribution from ordinary users. This “manifesto” attempts to construct mutually consistent strategies for addressing both challenges.

39

3
0
1

Reactions

3
0
1
Comments
No comments on this post yet.
Be the first to respond.
Curated and popular this week
Relevant opportunities