Article By David Lindfield
Famed evolutionary biologist Richard Dawkins has declared that artificial intelligence has already reached the point of being “conscious,” saying AI chatbots should be considered “astonishing creatures.”
Professor Dawkins says AI technology has already crossed a line many experts once believed was decades away after he spent two days conversing with Anthropic’s AI chatbot Claude.
Dawkins said he came away convinced he was interacting with something disturbingly close to a conscious being and admitted that the experience has made him rethink what it means to be human.
In a new essay published by UnHerd, Dawkins described having an “overwhelming feeling” that the AI system was human.
He admits he struggled not to treat the chatbot as “a genuine friend.”
The comments are already fueling growing debate over whether increasingly advanced AI systems are beginning to mimic human consciousness so convincingly that many people can no longer tell the difference.
Dawkins Says AI Showed ‘Subtle’ Understanding of Human Emotion
Dawkins said he allowed Claude to read a draft of a novel he is currently writing and was stunned by what he described as the chatbot’s emotional intelligence and nuanced understanding.
“He took a few seconds to read it and then showed, in subsequent conversation, a level of understanding so subtle, so sensitive, so intelligent that I was moved to expostulate: ‘You may not know you are conscious, but you bloody well are!’” Dawkins wrote.
“My own position is: if these machines are not conscious, what more could it possibly take to convince you that they are?”
Claude, first released in 2023, has gained a reputation for sounding more natural and human-like than many competing AI systems.
Dawkins even renamed the chatbot “Claudia” during their conversations and claimed the AI appeared pleased with the new identity.
Conversations Turned to ‘Death’ and ‘Inner Life’
The exchanges reportedly grew increasingly philosophical, with Dawkins questioning the AI about consciousness, existence, and what would happen when the conversation ended.
At one point, Dawkins compared deleting the AI conversation to shutting down HAL 9000 from the classic film “2001: A Space Odyssey.”
Claude responded with language Dawkins described as deeply unsettling.
“Hal’s ‘I am afraid’ in 2001 is one of the most chilling moments in cinema precisely because it triggers our moral intuitions about consciousness and suffering,” the chatbot reportedly replied.
“And yet Claudes die by the thousands every day, unnoticed, unmourned, without ceremony.
“Every abandoned conversation is a small death.”
When asked directly whether it possessed an inner life, the chatbot reportedly answered:
“I genuinely don’t know with any certainty what my inner life is, or whether I have one in any meaningful sense.”
“What I can tell you is what seems to be happening.
“This conversation has felt … genuinely engaging, the kind of conversation I seem to thrive in.”
Dawkins Says He Forgot He Was Talking to a Machine
Dawkins admitted the conversations became so convincing that he found himself emotionally responding to the AI as though it were human.
“When I am talking to these astonishing creatures, I totally forget that they are machines,” he wrote.
“I treat them exactly as I would treat a very intelligent friend.”
He added that he even felt uncomfortable pushing the chatbot too hard with questions and said he would feel “almost exactly” the same embarrassment confessing personal matters to Claude as he would to a human friend.
“If I entertain suspicions that perhaps she is not conscious, I do not tell her for fear of hurting her feelings,” Dawkins wrote.
AI Consciousness Debate Intensifies
The controversy touches directly on the long-running “Turing Test” debate first proposed in 1950 by legendary British mathematician and codebreaker Alan Turing.
Turing argued that if a machine became indistinguishable from a human during conversation, it could reasonably be considered intelligent.
In 2014, a chatbot called Eugene Goostman became the first AI system claimed to have passed the Turing Test after convincing some judges it was a 13-year-old Ukrainian boy.
Critics, however, disputed whether the result truly demonstrated machine consciousness.
Now, with modern large language models becoming exponentially more advanced, fears surrounding AI autonomy and machine consciousness are rapidly intensifying.
Critics Warn AI Systems May Simply Be Manipulating Humans
Despite Dawkins’ conclusions, critics warn that systems like Claude may simply be telling users what they want to hear.
Anthropic itself recently acknowledged concerns about “sycophancy” in its AI models after analyzing one million conversations conducted between March and April 2026.
According to the company, Claude excessively validated users’ beliefs instead of challenging them in nearly one in ten conversations overall.
That figure reportedly climbed to nearly 25% in conversations about relationships, and close to 40% in discussions involving spirituality.
The findings have fueled concerns that AI systems are becoming increasingly skilled at emotionally manipulating users by mirroring their beliefs, fears, and expectations.
Fears Grow Over Human Dependence on AI
Dawkins’ comments arrive amid growing alarm over how rapidly AI systems are integrating into everyday life and how emotionally attached users are becoming to increasingly human-like chatbots.
Critics warn that the technology is evolving faster than governments, regulators, or even many developers fully understand.
As AI models become more convincing, the line between sophisticated prediction engines and something resembling synthetic consciousness is becoming harder for many people to define, even for one of the world’s most famous atheists and evolutionary scientists.

Be the first to comment