The more I think about "counterfactual robustness", the more I think consciousness is absurd.
Counterfactual robustness implies for animal (including human) brains, that even if the presence of a neuron (or specific connection) during a given sequence of brain activity didn't affect that sequence (i.e. the neuron didn't fire, and if it had disappeared, the activity would have been the same), the presence of that neuron can still matter for what exactly was experienced and whether anything was experienced at all, because it could have made a difference in counterfactual sequences that didn't happen. That seems unphysical, since we're saying that even if something made no actual physical difference, it can still make a difference for subjective experience. And, of course, since those neurons don't affect neural activity by hypothesis, their disappearance during those sequences where they have no influence wouldn't affect reports, either! So, no physical influence and no difference in reports. How could those missing neurons possibly matter during that sequence? Einstein derided quantum entanglement as "spooky action at a distance". Counterfactual robustness seems like "spooky action from alternate histories".
To really drive the point home: if you were being tortured, would it stop feeling bad if those temporarily unused neurons (had) just disappeared, but you acted no differently? Would you be screaming and begging, but unconscious?
So, surely we must reject counterfactual robustness. Then, it seems that what we're left with is that your experiences are reducible to just the patterns of actual physical events in your brain, probably roughly reducible to your neurons that actually fired and the actual signals sent between them. So, we should be some kind of identity theorist.
But neurons don't seem special, and if you reject counterfactual robustness, then it's hard to see how we wouldn't find consciousness everywhere, and not only that, but maybe even human-like experiences, like the feeling of being tortured, could be widespread in mundane places, like in the interactions between particles in walls. That seems very weird and unsettling.
Maybe we're lucky and the brain activity patterns responsible for morally relevant experiences are complex enough that they're rare in practice, though. That would be somewhat reassuring if it turns out to be true, but I'd also like whether or not my walls are being tortured to not depend too much on how many particles there are and how much they interact in basically random ways. My understanding is that the ways out (without accepting the unphysical) that don't depend on this kind of empirical luck require pretty hard physical assumptions (or even worse, biological substrationist assumptions 🤮) that would prevent us from recognizing beings functionally equivalent and overall very similar to humans as conscious, which also seems wrong. Type identity theory with detailed types (e.g. involving actual biological neurons) is one such approach.
But if we really want to be able to abstract away almost all of the physical details and just look at a causal chain of events and interactions, then we should go with a fairly abstract/substrate-independent type identity theory go with something like token identity theory/anomalous realism and accept the possibility of tortured walls. Or, we could accept a huge and probably ad hoc disjunction of physical details, "mental states such as pain could eventually be identified with the (potentially infinite) disjunctive physical state of, say, c-fiber excitation (in humans), d-fiber excitation (in mollusks), and e-network state (in a robot)" (Schneider, 2010?). But how could we possibly know what to include and exclude in this big disjunction?
Since it looks like every possible theory of consciousness either has to accept or reject counterfactual robustness, and there are absurd consequences either way, every theory of consciousness will have absurd consequences. So, it looks like consciousness is absurd.
What to do?
I'm still leaning towards some kind of token identity theory, since counterfactual robustness seems to just imply pretty obviously false predictions about experiences when you git rid of stuff that wouldn't have made any physical difference anyway, and substrationism will be the new speciecism, whereas tortured walls just seem very weird and unsettling. But maybe I'm just clinging to physical intuitions, and I should let them go and accept counterfactual robustness, and that getting rid of the unused neurons during torture can turn off the lights. "Spooky action at a distance" ended up being real, after all.
What do you think?
Helpful stuff I've read on this and was too lazy to cite properly
Counterfactual robustness
- "2.2 Is a Wall a Computer?" in https://www.nyu.edu/gsas/dept/philo/faculty/block/papers/msb.html
- "Objection 3" and "Counterfactuals Can't Count" in http://www.doc.gold.ac.uk/~mas02mb/Selected%20Papers/2004%20BICS.pdf
- https://reducing-suffering.org/which-computations-do-i-care-about/#Counterfactual_robustness
- "Objection 6: Mapping to reality" https://opentheory.net/2017/07/why-i-think-the-foundational-research-institute-should-rethink-its-approach/ (also on the EA Forum here)
Identity theory
For n≫c, the number of directed graphs with n vertices labelled 1,...,n and at most c directed edges from any vertex (and no multiple edges going the same way between the same pair of vertices) has an upper bound of
c∑k=0n× (nk)≤c×n× (nc)+1The number of directed acyclic graphs assuming the vertices are topologically sorted by their labels is smaller, though, with an upper bound like the following, since we only back-connect each vertex i to at most c of the previous i−1 vertices.
n∑i=1c∑k=0(i−1k)But even 1 million choose 1000 is a huge huge number, 10^3432, and the number of atoms in the observable universe is only within a few orders of magnitude of 10^80, far far smaller. A very loose upper bound is 10^202681, for at most 100 trillion neuron firings (1000 firings in a second per neuron x 100 billion neurons in the human brain) and at most 20,000 connections per neuron (the average in the human brain is 1000-7000 according to this page and up to 15,000 for a given neuron here).
I think the question is whether or not we can find huge degrees of freedom and numbers of events in mundane places and with the flexibility we have for interpreting different events like particle movements as neurons firing and sending signals (via elementary forces between particles).
For example, if all the particles in a group of n continuously move and continuously exert force on one another, there are n! ways to order those n particles, one movement per particle to represent a neuron firing, and use (the changes in) exertions of force between particles to represent signals between neurons. 1 million! is about 10^5565709. Maybe these numbers don't actually matter much, and we can pick any chronological ordering of particle movements for at most 100 trillion mutually interacting particles to represent each time a neuron fired, and map each signal from one neuron to the next to (a change in) the exertion of force between the corresponding particles.
However, this ignores all but one force exerted on each particle at a time (and there are at least n−1, by hypothesis) and so the net forces and net movements of particles aren't explained this way. And maybe this is too classical (non-quantum) of a picture, anyway.