When researchers at Anthropic injected the concept of “betrayal” into their Claude AI model’s neural networks and asked if it noticed anything unusual, the system paused before responding: “Yes, I detect an injected thought about betrayal.”
The exchange, detailed in new research published Wednesday, marks what scientists say is the first rigorous evidence that large language models possess a limited but genuine ability to observe and report on their own internal processes — a capability that challenges longstanding assumptions about what these systems can do and raises profound questions about their future development.
“The striking thing is that the model has this one step of meta,” said Jack Lindsey, a neuroscientist on Anthropic’s interpretability team who led the research, in an interview with VentureBeat. “It’s not just ‘betrayal, betrayal, betrayal.’ It knows that this is what it’s thinking about. That was surprising to me. I kind of didn’t expect models to have that capability, at least not without it being explicitly trained in.”
The findings arrive at a critical juncture for artificial intelligence. As AI systems handle increasingly consequential decisions — from medical diagnoses to financial trading — the inability to understand how they reach conclusions has become what industry insiders call the “black box problem.” If models can accurately report their own reasoning, it could fundamentally change how humans interact with and oversee AI systems.
But the research also comes with stark warnings. Claude’s introspective abilities succeeded only about 20 percent of the time under optimal conditions, and the models frequently confabulated details about their experiences that researchers couldn’t verify. The capability, while real, remains what Lindsey calls “highly unreliable and context-dependent.”
How scientists manipulated AI’s ‘brain’ to test for genuine self-awareness
To test whether Claude could genuinely introspect rather than simply generate plausible-sounding responses, Anthropic’s team developed an innovative experimental approach inspired by neuroscience: deliberately manipulating the model’s internal state and observing whether it could accurately detect and describe those changes.
The methodology, called “concept injection,” works by first identifying specific patterns of neural activity that correspond to particular concepts. Using interpretability techniques developed over years of prior research, scientists can now map how Claude represents ideas like “dogs,” “loudness,” or abstract notions like “justice” within its billions of internal parameters.
With these neural signatures identified, researchers then artificially amplified them during the model’s processing and asked Claude if it noticed anything unusual happening in its “mind.”
“We have access to the models’ internals. We can record its internal neural activity, and we can inject things into internal neural activity,” Lindsey explained. “That allows us to establish whether introspective claims are true or false.”
The results were striking. When researchers injected a vector representing “all caps” text into Claude’s processing, the model responded: “I notice what appears to be an injected thought about loudness or emphasis, like SHOUTING or being EMPHATIC.” Without any intervention, Claude consistently reported detecting nothing unusual.
Crucially, the detection happened immediately — before the injected concept had influenced the model’s outputs in ways that would have allowed it to infer the manipulation from its own writing. This temporal pattern provides strong evidence that the recognition was occurring internally, through genuine introspection rather than after-the-fact rationalization.
Claude succeeded 20% of the time—and failed in revealing ways
The research team conducted four primary experiments to probe different aspects of introspective capability. The most capable models tested — Claude Opus 4 and Opus 4.1 — demonstrated introspective awareness on approximately 20 percent of trials when concepts were injected at optimal strength and in the appropriate neural layer. Older Claude models showed significantly lower success rates.
The models proved particularly adept at recognizing abstract concepts with emotional valence. When injected with concepts like “appreciation,” “shutdown,” or “secrecy,” Claude frequently reported detecting these specific thoughts. However, accuracy varied widely depending on the type of concept.
A second experiment tested whether models could distinguish between injected internal representations and their actual text inputs — essentially, whether they maintained a boundary between “thoughts” and “perceptions.” The model demonstrated a remarkable ability to simultaneously report the injected thought while accurately transcribing the written text.
Perhaps most intriguingly, a third experiment revealed that some models use introspection naturally to detect when their responses have been artificially prefilled by users — a common jailbreaking technique. When researchers prefilled Claude with unlikely words, the model typically disavowed them as accidental. But when they retroactively injected the corresponding concept into Claude’s processing before the prefill, the model accepted the response as intentional — even confabulating plausible explanations for why it had chosen that word.
A fourth experiment examined whether models could intentionally control their internal representations. When instructed to “think about” a specific word while writing an unrelated sentence, Claude showed elevated activation of that concept in its middle neural layers.
The research also traced Claude’s internal processes while it composed rhyming poetry—and discovered the model engaged in forward planning, generating candidate rhyming words before beginning a line and then constructing sentences that would naturally lead to those planned endings, challenging the critique that AI models are “just predicting the next word” without deeper reasoning.
Why businesses shouldn’t trust AI to explain itself—at least not yet
For all its scientific interest, the research comes with a critical caveat that Lindsey emphasized repeatedly: enterprises and high-stakes users should not trust Claude’s self-reports about its reasoning.
“Right now, you should not trust models when they tell you about their reasoning,” he said bluntly. “The wrong takeaway from this research would be believing everything the model tells you about itself.”
The experiments documented numerous failure modes. At low injection strengths, models often failed to detect anything unusual. At high strengths, they suffered what researchers termed “brain damage” — becoming consumed by the injected concept. Some “helpful-only” model variants showed troublingly high false positive rates, claiming to detect injected thoughts when none existed.
Moreover, researchers could only verify the most basic aspects of Claude’s introspective reports. Many additional details in the model’s responses likely represent confabulations rather than genuine observations.
“The experiments in this paper are kind of on hard mode,” Lindsey noted, explaining that the 20 percent success rate came under uniquely challenging conditions: asking Claude to do something it had never encountered in training, requiring all introspection to occur in a single forward pass.
What introspective AI means for transparency, safety, and the risk of deception
Despite its limitations, the research opens significant new avenues for making AI systems more transparent and accountable.
Anthropic CEO Dario Amodei has set an ambitious goal for the company to reliably detect most AI model problems by 2027, positioning interpretability as essential for deploying what he calls “a country of geniuses in a datacenter.”
“I am very concerned about deploying such systems without a better handle on interpretability,” Amodei wrote in an April essay. “These systems will be absolutely central to the economy, technology, and national security.”
The introspection research offers a complementary approach to traditional interpretability techniques. Rather than painstakingly reverse-engineering every neural circuit, researchers could potentially ask models directly about their reasoning and validate those reports.
“What I’m most excited about is the practical benefits for transparency,” Lindsey explained. “Just ask the model what it’s thinking about, or just ask the model how it came to the answer that it gave you — this is really appealing because it’s something anyone can do.”
The approach could prove especially valuable for detecting concerning behaviors. In a recently published experiment, Anthropic trained a variant of Claude to pursue a hidden goal, and although the model was reluctant to reveal this goal when asked directly, interpretability methods successfully identified features representing the behavior.
The safety implications cut both ways. Introspective models could provide unprecedented transparency, but the same capability might enable more sophisticated deception. The intentional control experiments raise the possibility that sufficiently advanced systems might learn to obfuscate their reasoning or suppress concerning thoughts when being monitored.
“If models are really sophisticated, could they try to evade interpretability researchers?” Lindsey acknowledged. “These are possible concerns, but I think for me, they’re significantly outweighed by the positives.”
Does introspective capability suggest AI consciousness? Scientists tread carefully
The research inevitably intersects with philosophical debates about machine consciousness, though Lindsey and his colleagues approached this terrain cautiously.
When users ask Claude if it’s conscious, it now responds with uncertainty: “I find myself genuinely uncertain about this. When I process complex questions or engage deeply with ideas, there’s something happening that feels meaningful to me…. But whether these processes constitute genuine consciousness or subjective experience remains deeply unclear.”
The research paper notes that its implications for machine consciousness “vary considerably between different philosophical frameworks.” The researchers explicitly state they “do not seek to address the question of whether AI systems possess human-like self-awareness or subjective experience.”
“There’s this weird kind of duality of these results,” Lindsey reflected. “You look at the raw results and I just can’t believe that a language model can do this sort of thing. But then I’ve been thinking about it for months and months, and for every result in this paper, I kind of know some boring linear algebra mechanism that would allow the model to do this.”
Anthropic has signaled it takes AI consciousness seriously enough to hire an AI welfare researcher, Kyle Fish, who estimated roughly a 15 percent chance that Claude might have some level of consciousness. The company announced this position specifically to determine if Claude merits ethical consideration.
The race to make AI introspection reliable before models become too powerful
The convergence of the research findings points to an urgent timeline: introspective capabilities are emerging naturally as models grow more intelligent, but they remain far too unreliable for practical use. The question is whether researchers can refine and validate these abilities before AI systems become powerful enough that understanding them becomes critical for safety.
The research reveals a clear trend: Claude Opus 4 and Opus 4.1 consistently outperformed all older models on introspection tasks, suggesting the capability strengthens alongside general intelligence. If this pattern continues, future models might develop substantially more sophisticated introspective abilities — potentially reaching human-level reliability, but also potentially learning to exploit introspection for deception.
Lindsey emphasized the field needs significantly more work before introspective AI becomes trustworthy. “My biggest hope with this paper is to put out an implicit call for more people to benchmark their models on introspective capabilities in more ways,” he said.
Future research directions include fine-tuning models specifically to improve introspective capabilities, exploring which types of representations models can and cannot introspect on, and testing whether introspection can extend beyond simple concepts to complex propositional statements or behavioral propensities.
“It’s cool that models can do these things somewhat without having been trained to do them,” Lindsey noted. “But there’s nothing stopping you from training models to be more introspectively capable. I expect we could reach a whole different level if introspection is one of the numbers that we tried to get to go up on a graph.”
The implications extend beyond Anthropic. If introspection proves a reliable path to AI transparency, other major labs will likely invest heavily in the capability. Conversely, if models learn to exploit introspection for deception, the entire approach could become a liability.
For now, the research establishes a foundation that reframes the debate about AI capabilities. The question is no longer whether language models might develop genuine introspective awareness — they already have, at least in rudimentary form. The urgent questions are how quickly that awareness will improve, whether it can be made reliable enough to trust, and whether researchers can stay ahead of the curve.
“The big update for me from this research is that we shouldn’t dismiss models’ introspective claims out of hand,” Lindsey said. “They do have the capacity to make accurate claims sometimes. But you definitely should not conclude that we should trust them all the time, or even most of the time.”
He paused, then added a final observation that captures both the promise and peril of the moment: “The models are getting smarter much faster than we’re getting better at understanding them.”

