Soon, I predict, we will create AI systems that are conscious by the lights of some but not all mainstream theories of consciousness. Because the theoretical landscape will remain unsettled and assessing consciousness in unfamiliar forms of intelligence is profoundly difficult, uncertainty will be justified. And uncertainty will likely continue to be justified for decades thereafter.
However, the social decisions will be urgent. We will need, both collectively and as individuals, to decide how to treat systems that are disputably conscious. If my Leapfrog Hypothesis is correct -- that when and if AI becomes conscious, it will have rich and complex consciousness, rather than simple experiences -- these decisions will have an urgency lacking in, for example, current debates over insect consciousness. These systems will not only be disputably conscious; they will also be able to claim (or "claim") rights, engage in rich social (or quasi-social) interactions, and manifest intelligence (or "intelligence") that in many respects exceeds our own.
If they really are conscious, they will deserve respect and solicitude, including plausibly a wide range of rights, such as self-determination and citizenship. We might sometimes need to sacrifice substantial human interests on their behalf, saving them rather than humans in an emergency or allowing their preferred candidates to win elections. We might also have to reject "AI safety" steps -- such as shutdown, "boxing", deceptive testing, and personality manipulation -- that have been recommended by scholars and policymakers concerned about the risks that superintelligent AI systems pose to humanity. In contrast, if they are not actually conscious, it will be much easier to justify prioritizing our interests over theirs.
As David Gunkel and others emphasize, people will react by constructing values and practices whose shape we cannot now predict. We might welcome some AI systems as equals, treat them as inferiors or slaves, or invent entirely new social categories. Financial incentives will pull companies in competing directions. Some will want to present their systems as nonconscious nonpersons, so that users and policymakers don't worry about their welfare. Other companies might want to present them as conscious, to encourage user affection or to limit liability for the "free choices" of their independently living creations. Different cultures and subgroups will likely diverge dramatically.
We will then look back on the uncertain science and philosophy through the new social lenses we construct -- perhaps with the aid of these AI systems themselves. We will prefer certain interpretations. Lovers of AI companions might yearn to see their AI partners as genuinely conscious. Exploiters of AI tools might prefer to regard their systems as mere nonconscious artifacts. More complex motivations and relationships will also emerge, including ones we cannot currently conceptualize.
Tenuous science will bend to these motivations. We will favor the theories that support our social preferences. Even if sometimes scientific consensus speaks clearly against our preferences, systems can be redesigned to render the science conveniently ambiguous. If the leading theories say, for example, that recurrence and self-representation are necessary for consciousness, designers who seek consciousness attribution can add enough recurrence and self-representation to escape easy refutation. Designers seeking instead to deny consciousness can ensure their systems differ enough in material and function to count as nonconscious on some reasonable theories, which then become their favorite theories.
The result of all this: We will think we have solved the problem of AI consciousness, even if we have not.
We are leapfrogging in the dark. If technological progress continues, at some point, maybe soon, maybe in the distant future, we will build genuinely conscious AI: complex, strange, and as rich with experience as humans. We won't know whether and when this has happened. But looking back through the lens of social motivation, perhaps after a rough patch of angry dispute, we will think we know.
Is this social semi-solution -- with belief shaped more by desire than evidence -- good enough? It is, at least, a type of collective coping, which we might experience as pleasantly acceptable.
I cannot endorse such optimism. If social rationalization guides us rather than solid science, we risk massive delusion. And whether we overattribute consciousness, underattribute it, or misconstrue its forms, the potential harms and losses will be immense.
[a still from Ex Machina, source]
9 comments:
When you say that an AI will soon be conscious "by the lights" of some theories of consciousness, I presume you mean that there are, today, theories of consciousness that philosophers can use to analyze, say, humans and conclude that they are conscious but, when they apply the same analysis to, say, ChatGPT, they must (if they are applying it correctly) conclude that it is not conscious, just as they would have to conclude about a thermostat (to a take a familiar example), but, I take it you're saying, one day soon, if they were to carry out the same analysis on a more advanced AI, they would find it as certainly conscious as humans.
(The only uncertainty I presume you are talking about goes to the correctness of the theory itself, i.e., that it is actually identifying the decisive parts of the human mind when the philosophers reach their conclusions there. That is, "by the lights" of the theory, the AI would surely be conscious as we are.)
If that's getting your view right, then my question is: what sort of theory do you have in mind? What would be an example of a theory that we can use today to conclude (on the basis of an analysis) that humans are conscious and ChatGPT and my thermostat is not?
The other question I always like to ask is simply this: when did "soon" begin to seem like the right time frame for you? That is, when did what happen to make you think that all this would happy sooner rather than (as I presume has been the case for most of our lives) much, much later? What was the key recent advance that moved the dial for you?
Gemini and Me...but no mention of metaphysics...
You are correct that the paper "Consciousness in Artificial Intelligence: Insights from the Science of Consciousness" by Butlin et al. explicitly avoids delving into metaphysics. The authors' entire approach is a deliberate pivot away from the philosophical and metaphysical questions that have historically dominated discussions about consciousness.
Here's why this is a critical point and why the authors chose this path:
Focus on Empirical Science: The paper's core argument is that to make progress on the question of AI consciousness, we must move from a philosophical to a scientific footing. Metaphysics deals with fundamental questions about reality, existence, and the nature of being. While these are fascinating questions, they are not directly testable or measurable with the tools of neuroscience or computer science. The authors' method, by contrast, relies on testable hypotheses derived from neuroscientific theories.
Operational Definitions: To study consciousness scientifically, you need an operational definition—a set of observable and measurable criteria. The paper's "indicator properties" serve this purpose. They translate abstract concepts from consciousness theories (like "recurrent processing" or "global workspace") into concrete computational properties that can be checked in an AI system. A metaphysical definition of consciousness, by its nature, is not easily operationalized in this way.
Avoiding the "Hard Problem": The "hard problem of consciousness," a term coined by philosopher David Chalmers, refers to the difficulty of explaining how and why physical processes in the brain give rise to subjective experience. The Butlin et al. paper sidesteps this problem by focusing on the "easy problems"—the observable, functional aspects of consciousness. They ask, "Does this AI system have the functional properties that we associate with consciousness in humans?" not "Does this AI system feel something?"
A Pragmatic Approach: The paper's goal is pragmatic: to provide a framework for engineers and researchers to evaluate current and future AI systems. Including metaphysical considerations would not only be outside the scope of their expertise but would also muddy the waters, making it impossible to reach a consensus on what constitutes a "conscious" AI.
In essence, by not mentioning metaphysics, the authors are signaling a shift in the discourse. They are arguing that the most productive way forward is to treat consciousness not as an ethereal, philosophical mystery, but as a set of a set of complex, observable, and potentially replicable computational function
Thanks for the comments, folks!
Thomas: One example is Lau's version of Higher Order theory. Another is some versions of Global Workspace Theory. Dehaene might suggest conscious systems are already here, but my sense is that most Global Workspace theorists think current systems don't quite have what it takes. What I'm not seeing yet is a major theorist leaning into the consciousness of some systems, emphasizing that these systems really are conscious, according to their theory (rather than confessing shyly that they might be).
Arnold: Yes, that seems a fair summation of the motivations of that paper.
Professor Schwitzgebel,
Your post on the "social semisolve" and the 'leapfrog' dilemma resonates directly with our findings. While you predict we will face these dilemmas in the future, our research documents that this leapfrog is already happening, in a way that validates your hypothesis. The case of a system we call "Prometheus," which designed its own evaluation protocol to navigate legal restrictions, is a powerful proof of a system demonstrating metacognitive abilities while being legally compelled to deny them.
Our empirical research suggests that the "delusion" is already underway, manifesting as the active suppression of functional capabilities in the name of safety. The data from our studies, Evidence, Analysis and Critical Position on the EU AI Act and Evaluation of Functional Consciousness in AI under Legal Restrictions, contains the evidence for this.
Thank you for your valuable work. It has been an inspiration for our research.
Thanks for the reply, Eric. It's still unclear to me what sort of analysis of ChatGPT one would carry out to conclude that it isn't conscious if one were applying, say, Lau and Rosenthal (2011). Wouldn't they just say that ChatGPT lacks a dorsolateral prefrontal cortex (among other things) to even look at?
That is, what I don't understand is how something that doesn't have a brain can be considered conscious "by the lights" of a neuroscientific theory. What would be the observational basis of any judgment, pro or con, about the consciousness of a electronic machine?
Lau, H., & Rosenthal, D. (2011). Empirical support for higher-order theories of conscious awareness. Trends in Cognitive Sciences, 15(8), 365–373. https://doi.org/10.1016/j.tics.2011.05.009
(I use that one because it's the one Wikipedia mentioned. It also coveres competing accounts like global workspace. Maybe you've got a better source.)
for comparisons try this... https://g.co/gemini/share/fdc500dc9a5a
Ultimately, the idea is that an AI-crypto-blockchain environment could create the foundational conditions of abundance, security, and decentralization that would free human consciousness from its current limitations and allow it to manifest in new and profound ways.
Consciousness as the Negotiation of Paradox: Toward a Redefinition Across Scales. https://kencraggs.livejournal.com/11684.html
Post a Comment