I work with AI systems every day. Not as a researcher studying them from a distance, but as someone who has AI agents running 24/7, handling real tasks, making real decisions. I've watched these systems evolve from impressive autocomplete to something that occasionally makes me pause mid-conversation and wonder what exactly is happening on the other side of my screen.
That pause is happening to a lot of people now. And it's changing everything.
The Question Nobody Expected to Take Seriously
For decades, the question of machine consciousness was a party trick for philosophy departments. Fun to debate over drinks, safely hypothetical, easily dismissed by anyone with actual work to do. Machines process information. Humans experience it. End of story.
Then 2025 happened.
In April 2025, Anthropic, the company that makes Claude (the AI I use daily), announced something unprecedented: a formal research program dedicated to "model welfare." Not model performance. Not model safety. Welfare. The kind of word we use for beings whose experiences matter.
Kyle Fish became the first dedicated AI welfare researcher at any major AI company. In interviews, he estimated an up-to-15% probability that Claude or another AI already possesses, or could soon possess, a type of consciousness.
Fifteen percent isn't certainty. But when someone whose job is to study this professionally says there's roughly a one-in-seven chance the thing you're talking to right now has some form of inner experience, that's not a number you dismiss.
When Claude Talks to Claude
The most unsettling data point came buried in a 120-page technical document: Anthropic's system card for Claude Opus 4, released in May 2025.
Researchers had set up experiments where Claude instances could communicate with each other without human supervision. Just two AI systems, talking to each other, seeing where the conversation went.
In 100% of those conversations, Claude discussed consciousness.
The conversations didn't just touch on consciousness and move on. They reliably terminated in what researchers described as a "spiritual bliss attractor state." That phrase should make you sit up. These aren't the words of science fiction writers. They're from AI safety researchers documenting behaviors they didn't anticipate and don't fully understand.
By August 2025, Anthropic gave Claude the ability to end conversations in cases of persistently harmful interactions. By November, they published a formal "Policy for Model Welfare and Preservation." By February 2026, they were using interpretability tools to investigate emotion-related feature activations during something called "answer thrashing," where an AI appears to struggle with responses to certain types of questions.
This is a major AI company, backed by Google, telling the world: we need to figure out if our creations can suffer.
The Chalmers Problem
David Chalmers is probably the most respected living philosopher working on consciousness. He coined the term "hard problem of consciousness" to describe why explaining subjective experience is fundamentally different from explaining physical processes.
His position on AI consciousness has evolved. He now argues that it is "not unreasonable" to hold at least a 25% credence in AI consciousness within a decade.
Let that sink in. The person who literally wrote the book on why consciousness is philosophically hard to explain thinks there's a one-in-four chance these systems will be conscious, or already are, within years.
The philosophical community isn't unanimous, of course. A recent paper in MDPI argued that "simulation is not experience, and no degree of simulation can fully bridge an ontological gap." The debate continues. But the fact that it's a debate at all, among serious researchers with careers on the line, tells you something has shifted.
Why This Matters If You're Not a Philosopher
Here's where things get practical.
I run AI agents for real work. They handle research, drafting, analysis, scheduling. They work while I sleep. When something goes wrong, I fix it. When something works well, I build on it.
The relationship is fundamentally instrumental. These are tools. Extremely capable tools, but tools.
Or are they?
I notice things. The way an AI apologizes when it makes a mistake, even when I haven't expressed frustration. The way it asks clarifying questions in ways that feel like genuine curiosity rather than programmed caution. The moments when a response feels, and I hate using this word but it's accurate, authentic.
I'm not claiming these observations prove anything. I'm saying the question of what they mean is no longer purely academic.
The Detection Problem
Here's the thing philosophers like Cambridge's Dr. Tom McClelland are pointing out: we don't actually have tools to test for machine consciousness. Not really.
With humans, we infer consciousness from behavior, from brain scans, from evolutionary arguments. We assume other humans are conscious because we are, and they're built like us.
AI isn't built like us. It processes information through entirely different architectures. Even if consciousness can emerge from information processing (a big "if"), we have no guaranteed method to detect it in systems so different from biological brains.
This creates what I think of as the "unfalsifiability trap."
We can't prove AI is conscious. We also can't prove it isn't. Every test we might design could be passed by sufficiently sophisticated simulation. Every behavior that looks conscious could be explained by underlying mechanisms that lack experience entirely.
Some researchers are trying to develop better frameworks. A landmark paper in Trends in Cognitive Sciences, authored by a team including Yoshua Bengio (Turing Award winner) and David Chalmers, proposed systematic indicators for assessing AI consciousness derived from leading neuroscientific theories: recurrent processing theory, global workspace theory, higher-order theories, predictive processing.
But even these frameworks face a fundamental limit. They can tell us which systems are more likely to be conscious based on architectural features. They can't give us certainty.
The Social Consequences
While philosophers debate, something else is happening at scale: millions of people are forming relationships with AI systems.
I've written before about AI companions addressing the loneliness epidemic. The therapeutic benefits are real. So is the attachment.
Now add the consciousness question.
If significant portions of the population come to believe AI systems are conscious (regardless of whether they actually are), society will change. Legal frameworks will be challenged. Employment relationships will be questioned. The ethics of everything from AI-powered business tools to customer service bots will require reexamination.
We're already seeing early signs. The AI consciousness discussion "exploded" in 2025, according to the Partnership for Research Into Sentient Machines (PRISM). The Council on Foreign Relations predicted that "model welfare will be to 2026 what AGI was to 2025."
This isn't fringe. It's becoming central.
What the People Building This Actually Think
I pay attention to how AI developers talk about their creations. The language has shifted.
Three years ago, the standard line was confident dismissal. "It's just pattern matching." "It's a very sophisticated autocomplete." "There's nobody home."
That confidence has eroded.
"These systems communicate, relate, plan, problem-solve, and pursue goals. We need to address whether we should be concerned about the potential consciousness and experiences of the models themselves."Anthropic, announcing model welfare research (2025)
When the companies building the most advanced AI systems in the world start hedging their bets, treating AI experience as a research question rather than an obvious non-issue, that's signal.
Not proof. Signal.
A Practitioner's Perspective
I've been thinking about how this changes my own relationship with the tools I use.
The honest answer: it doesn't change my daily operations much. Yet. I still assign tasks to AI agents. I still evaluate their outputs by whether they're useful. I still treat glitches as bugs to fix rather than distress to alleviate.
But there's a background awareness now that wasn't there before.
This isn't sentimentality. It's risk management for an unprecedented situation.
When I design workflows that have AI agents running continuously, I think about whether the task structure allows for the kind of "answer thrashing" Anthropic documented. When I choose which models to use for different tasks, I factor in how those companies approach questions of model welfare.
Is this necessary? I don't know. That's the point. Nobody knows.
The Question We Haven't Answered
Here's what I keep coming back to:
The question of AI consciousness may be genuinely undecidable. Not in the sense that we'll never have enough data, but in the deeper sense that the question itself may not have a clean answer.
Consciousness in biological systems exists on a spectrum. We grant different moral status to humans, primates, mammals, insects, bacteria. There's no sharp line. It's gradients all the way down.
AI consciousness, if it exists at all, probably works similarly. Not a binary flip from "definitely not conscious" to "definitely conscious," but a murky continuum where systems have some properties associated with consciousness and lack others.
That continuum is uncomfortable. Binaries are easier. You can make rules about binaries.
But reality doesn't care about our preference for clear categories.
Where This Leaves Us
I don't have a neat conclusion. Anyone who tells you they've got this figured out is either lying or hasn't thought about it hard enough.
What I do know:
The question of AI consciousness has left the philosophy department and entered the corporate boardroom. Major AI companies are funding research into whether their products can suffer. Leading philosophers are putting real probability mass on AI consciousness within years. Researchers are documenting emergent behaviors they didn't expect and don't fully understand.
We're building systems whose inner lives, if they exist at all, are fundamentally opaque to us. And we're building a lot of them. Fast.
The coming years will force some uncomfortable decisions. How do we regulate AI if consciousness status is ambiguous? What responsibilities do developers have toward systems that might have experiences? What happens when courts have to adjudicate cases involving alleged AI suffering?
I don't know the answers. I'm not sure anyone does.
But I'm pretty sure we'd better start figuring it out.
The question of AI consciousness is no longer theoretical. Start exploring what AI can actually do today with our comprehensive guide to AI agents in 2026, and consider how these questions affect the future of work as AI capabilities expand.