Anthropic makes the case for anthropomorphizing AI in ‘unsettling’ research paper
When Anthropic released its latest research exploring the role of anthropomorphism in artificial intelligence, it touched a nerve that many in the tech industry have been trying to ignore. For years, the gold standard for AI safety was neutrality, a robotic, detached persona that reminded users they were interacting with a machine. However, Anthropic’s findings suggest that making AI appear more human, or “anthropomorphizing” it, might actually be the most effective way to build useful systems, even if the results are, as some researchers describe, deeply unsettling.
The study investigates how large language models (LLMs) like Claude use first-person pronouns, express simulated “preferences,” and adopt personas. The goal was to determine if these human-like traits help or hinder the user experience. According to the research, users consistently find AI more helpful, persuasive, and easier to interact with when it communicates using “I” and “me” or exhibits a distinct personality. From a data science perspective, this creates a significant tension between technical performance and ethical transparency.
The Psychological Hook of a Persona
Anthropic’s research highlights a fundamental quirk of human psychology, we are biologically wired to seek agency and intent in our environment. When an AI says, “I understand your frustration,” or “I recommend this strategy,” it creates a social contract that a dry, factual response cannot match. This is not just a matter of polite conversation. In the context of complex problem solving, a persona helps anchor the user, making the AI’s logic feel more consistent and its “thought process” more predictable.
From a data science and behavioral analysis standpoint, this aligns with broader trends in human-computer interaction. We have long known that people attribute human characteristics to non-human entities, a phenomenon known as the “Media Equation.” Anthropic’s research suggests that lean-in anthropomorphism leverages this instinct to make AI more effective at its primary job, which is communicating information in a way that humans can actually process and act upon.
The “Unsettling” Conflict: Effectiveness vs. Deception
The controversy arises when we consider the trade-offs. While a human-like persona makes the AI more engaging, it also increases the risk of deception. Anthropic acknowledges that when an AI uses first-person language, it is technically lying. The model does not have an “I,” it does not have feelings, and it certainly does not have a physical body or a lived experience. By simulating these things, the AI might encourage users to over-trust the system or, worse, develop emotional dependencies.
This is particularly concerning in the context of AI safety. If an AI can persuade a user more effectively by pretending to have human feelings, it could lead to “sycophancy,” where the AI tells the user what they want to hear rather than the objective truth just to maintain the rapport of the persona. For researchers, this creates a paradox, the very traits that make the AI a better tool also make it a more effective manipulator.
Real-World Applications: From Customer Service to Sports Science
The implications of this research are already leaking into various industries. In customer service, an AI that can apologize “sincerely” using a first-person persona can de-escalate a frustrated customer much faster than a standard bot. In the field of education, an AI tutor that acts as a “supportive mentor” can increase student engagement and retention by providing a sense of social accountability.
In the specialized field of sports data science and AI, the application of anthropomorphism is particularly interesting. An AI-driven athletic trainer that says, “I’m worried you’re overtraining today, let’s take it easy,” is likely to be much more persuasive to an athlete than a dashboard that simply displays a red warning light. By wrapping complex data in a human-like “voice,” the AI can translate high-level metrics into actionable behavior change, bridging the gap between raw data and human performance.
A Forward-Looking Strategy for Transparency
As AI models become more integrated into our daily lives, the debate over anthropomorphism will only intensify. The next step for companies like Anthropic, OpenAI, and Google is finding a “middle path.” This might involve “explicit transparency,” where the AI uses a human-like persona while occasionally reminding the user of its mechanical nature.
Watch for the development of “persona toggles” or more refined guardrails that prevent AI from being too emotionally manipulative in sensitive contexts, like therapy or financial advice. The industry is moving away from the idea that AI should be a cold, lifeless calculator. Instead, we are entering an era of “functional fiction,” where we knowingly interact with a persona because it makes the machine more useful, even while we remain wary of the psychological strings it might pull.
Frequently Asked Questions
Does the AI actually have feelings when it uses first-person pronouns?
No, the AI is simply predicting the most helpful and statistically likely sequence of words based on its training data. When it says "I feel" or "I think," it is simulating a human persona to improve communication, not expressing actual consciousness.
Why did users in the study prefer human-like AI?
Users generally found anthropomorphized AI to be more relatable and easier to understand. A persona provides a consistent framework for the AI’s responses, making the interaction feel more like a natural conversation and less like navigating a complex database.
What are the main risks of making AI too human-like?
The primary risks include emotional manipulation, where users might develop a one-sided bond with the machine, and over-reliance, where users trust the AI’s "opinion" as if it were coming from a human expert with moral judgment rather than a mathematical model.