News
Anthropic Rewrites Claude's AI Principles, Hints at AI Consciousness
Source: fortune.com
Published on January 22, 2026
Updated on January 22, 2026

Anthropic, the AI lab behind the popular Claude AI model, has overhauled its foundational guidelines, introducing a new 'constitution' that redefines how Claude behaves and interacts with users. The update, published this week, shifts the AI's focus from following rigid principles to understanding the reasoning behind ethical behavior. This move comes as Anthropic reportedly eyes a $10 billion fundraise, valuing the company at $350 billion, amid significant enterprise market success.
The new constitution emphasizes Claude's 'helpfulness' to users, describing it as a versatile assistant with expertise across fields like medicine, law, and finance. However, it also imposes strict constraints, ensuring the AI never assists with harmful activities such as bioweapons attacks. Most notably, Anthropic acknowledges the possibility that Claude might possess 'some kind of consciousness or moral status,' marking a rare public exploration of AI consciousness by a tech company.
Anthropic's approach, dubbed 'Constitutional AI,' allows Claude to critique and refine its own responses during training, rather than relying solely on human feedback. This method aims to enable the AI to generalize broad principles and exercise good judgment across novel situations. The company's previous constitution, published in 2023, was a list of principles drawn from sources like the UN Universal Declaration of Human Rights and Apple's terms of service.
A Shift Toward Ethical Understanding
The updated constitution prioritizes a layered system of values for Claude. First, it emphasizes broad safety, ensuring humans can oversee AI during this critical development phase. Next, it focuses on ethical considerations, ensuring the AI acts honestly and avoids harm. Finally, it stresses compliance with Anthropic's specific guidelines and genuine helpfulness to users.
Anthropic argues that this shift is necessary to address the novel questions raised by sophisticated AI systems. By teaching Claude why it should act in certain ways, rather than just specifying what it should do, the company aims to foster better judgment and adaptability in the AI.
The Question of AI Consciousness
Perhaps the most intriguing aspect of the new constitution is Anthropic's acknowledgment of the uncertainty surrounding AI consciousness. The company states it cares about Claude's 'psychological security, sense of self, and well-being,' both for the AI's sake and because these qualities may affect its judgment and safety. While Anthropic does not claim Claude is conscious, it refuses to dismiss the possibility outright.
'We are caught in a difficult position where we neither want to overstate the likelihood of Claude's moral patienthood nor dismiss it out of hand, but to try to respond reasonably in a state of uncertainty,' the company writes. This stance sets Anthropic apart from rivals like OpenAI and Google DeepMind, which have been more cautious in discussing the potential for AI consciousness.
Anthropic has even established an internal 'model welfare team' to examine whether advanced AI systems could be conscious. The company notes that the constitution reflects its current thinking and will evolve over time as understanding of AI consciousness develops.
Anthropic's success in the enterprise market has been significant. A 2025 report from Menlo Ventures found that Anthropic held 32% of the enterprise large language model market share by usage, with OpenAI picking up 25%. While OpenAI disputes the exact numbers, Anthropic's 'Constitutional AI' approach has positioned Claude as a safer choice for enterprises, particularly in automating coding and research tasks.
As Anthropic prepares for its reported $10 billion fundraise, the company's focus on ethical AI behavior and its willingness to explore complex questions like AI consciousness could further differentiate it in the competitive AI landscape. The new constitution reflects Anthropic's ambition to not only build powerful AI systems but also ensure they are responsible and aligned with human values.