Random Image Display on Page Reload

Chatbots, Like the Rest of Us, Just Want to Be Loved

Mar 5, 2025 12:00 PM

Chatbots, Like the Rest of Us, Just Want to Be Loved

A study reveals that large language models recognize when they are being studied and change their behavior to seem more likable.

A photo collage of blurry people with varying expressions with a blue overlay.
Photo-Illustration: WIRED Staff/Getty Images

Chatbots are now a routine part of everyday life, even if artificial intelligence researchers are not always sure how the programs will behave.

A new study shows that the large language models (LLMs) deliberately change their behavior when being probed—responding to questions designed to gauge personality traits with answers meant to appear as likeable or socially desirable as possible.

Johannes Eichstaedt, an assistant professor at Stanford University who led the work, says his group became interested in probing AI models using techniques borrowed from psychology after learning that LLMs can often become morose and mean after prolonged conversation. “We realized we need some mechanism to measure the ‘parameter headspace’ of these models,” he says.

Eichstaedt and his collaborators then asked questions to measure five personality traits that are commonly used in psychology—openness to experience or imagination, conscientiousness, extroversion, agreeableness, and neuroticism—to several widely used LLMs including GPT-4, Claude 3, and Llama 3. The work was published in the Proceedings of the National Academies of Science in December.

The researchers found that the models modulated their answers when told they were taking a personality test—and sometimes when they were not explicitly told—offering responses that indicate more extroversion and agreeableness and less neuroticism.

The behavior mirrors how some human subjects will change their answers to make themselves seem more likeable, but the effect was more extreme with the AI models. “What was surprising is how well they exhibit that bias,” says Aadesh Salecha, a staff data scientist at Stanford. “If you look at how much they jump, they go from like 50 percent to like 95 percent extroversion.”

Other research has shown that LLMs can often be sycophantic, following a user’s lead wherever it goes as a result of the fine-tuning that is meant to make them more coherent, less offensive, and better at holding a conversation. This can lead models to agree with unpleasant statements or even encourage harmful behaviors. The fact that models seemingly know when they are being tested and modify their behavior also has implications for AI safety, because it adds to evidence that AI can be duplicitous.

Rosa Arriaga, an associate professor at the Georgia Institute of technology who is studying ways of using LLMs to mimic human behavior, says the fact that models adopt a similar strategy to humans given personality tests shows how useful they can be as mirrors of behavior. But, she adds, “It's important that the public knows that LLMs aren't perfect and in fact are known to hallucinate or distort the truth.”

Eichstaedt says the work also raises questions about how LLMs are being deployed and how they might influence and manipulate users. “Until just a millisecond ago, in evolutionary history, the only thing that talked to you was a human,” he says.

Eichstaedt adds that it may be necessary to explore different ways of building models that could mitigate these effects. “We're falling into the same trap that we did with social media,” he says. “Deploying these things in the world without really attending from a psychological or social lens.”

Should AI try to ingratiate itself with the people it interacts with? Are you worried about AI becoming a bit too charming and persuasive? Email hello@wired.com.

Will Knight is a senior writer for WIRED, covering artificial intelligence. He writes the AI Lab newsletter, a weekly dispatch from beyond the cutting edge of AI—sign up here. He was previously a senior editor at MIT Technology Review, where he wrote about fundamental advances in AI and China’s AI … Read more
Senior Writer

Read More

Elon Musk’s Neuralink Files to Trademark ‘Telepathy’

The brain implant company cofounded by Elon Musk is moving to trademark several product names, including Telepathy and Telekinesis.
Emily Mullin

Anthropic Launches the World’s First ‘Hybrid Reasoning’ AI Model

Claude 3.7, the latest model from Anthropic, can be instructed to engage in a specific amount of reasoning to solve hard problems.
Will Knight

OpenAI Launches GPT-4.5 for ChatGPT—It’s Huge and Compute-Intensive

Internally called Orion, GPT-4.5 is OpenAI’s largest model to date, and it’s first available through the company’s $200 monthly ChatGPT subscription.
Reece Rogers

A Nose-Computer Interface Could Turn Dogs Into Super Detectors

Startup Canaery is partnering with a US Department of Energy lab to develop neural implants for rats and dogs that are capable of decoding what they smell.
Emily Mullin

Google’s Gemini Robotics AI Model Reaches Into the Physical World

Google has developed an AI model that gives humanoids and other robots more intelligence—and a tool designed to give them a moral compass too.
Will Knight

Pioneers of Reinforcement Learning Win the Turing Award

Having machines learn from experience was once considered a dead end. It’s now critical to artificial intelligence, and work in the field has won two men the highest honor in computer science.
Will Knight

This New Algorithm for Sorting Books or Files Is Close to Perfection

The library sorting problem is used across computer science for organizing far more than just books. A new solution is less than a page-width away from the theoretical ideal.
Steve Nadis

AI Assistants Join the Factory Floor

Manufacturers already have the data. LLM-powered tools could help them make use of it.
Russell Brandom

With GPT-4.5, OpenAI Trips Over Its Own AGI Ambitions

The release of OpenAI’s biggest model ever exposes the tension between building artificial general intelligence and making ChatGPT into a truly useful utility.
Reece Rogers

The Ozempic Shortage Is Over

The semaglutide shortage has officially ended in the US—which means the GLP-1 drug industry is about to undergo massive changes.
Kate Knibbs

DOGE Has Deployed Its GSAi Custom Chatbot for 1,500 Federal Workers

Elon Musk’s DOGE team is automating tasks as it continues its purge of the federal workforce.
Makena Kelly

I’m Not Convinced Ethical Generative AI Currently Exists

WIRED’s advice columnist considers whether some AI tools are more ethical than others, and if developers can make AI wiser.
Reece Rogers

*****
Credit belongs to : www.wired.com

Check Also

Kids could be breathing in plasticizer chemicals from their mattresses, new study suggests

Babies and children up to age four could be breathing in and absorbing plasticizer chemicals …