Random Image Display on Page Reload

The Chatbots Are Now Talking to Each Other

The Chatbots Are Now Talking to Each Other

ChatGPT-style chatbots that pretend to be people are being used to help companies develop new product and marketing ideas.

Smiling mouths on green backdrop

Photograph: Carol Yepes/Getty Images

Lena Anderson isn’t a soccer fan, but she does spend a lot of time ferrying her kids between soccer practices and competitive games.

“I may not pull out a foam finger and painted face, but soccer does have a place in my life,” says the soccer mom—who also happens to be completely made up. Anderson is a fictional personality played by artificial intelligence software like that powering ChatGPT.

Anderson doesn’t let her imaginary status get in the way of her opinions, though, and comes complete with a detailed backstory. In a wide-ranging conversation with a human interlocutor, the bot says that it has a 7-year-old son who is a fan of the New England Revolution and loves going to home games at Gillette Stadium in Massachusetts. Anderson claims to think the sport is a wonderful way for kids to stay active and make new friends.

In another conversation, two more AI characters, Jason Smith and Ashley Thompson, talk to one another about ways that Major League Soccer (MLS) might reach new audiences. Smith suggests a mobile app with an augmented reality feature showing different views of games. Thompson adds that the app could include “gamification” that lets players earn points as they watch.

The three bots are among scores of AI characters that have been developed by Fantasy, a New York company that helps businesses such as LG, Ford, Spotify, and Google dream up and test new product ideas. Fantasy calls its bots synthetic humans and says they can help clients learn about audiences, think through product concepts, and even generate new ideas, like the soccer app.

"The technology is truly incredible," says Cole Sletten, VP of digital experience at the MLS. “We’re already seeing huge value and this is just the beginning.”

Video: Fantasy

Fantasy uses the kind of machine learning technology that powers chatbots like OpenAI’s ChatGPT and Google’s Bard to create its synthetic humans. The company gives each agent dozens of characteristics drawn from ethnographic research on real people, feeding them into commercial large language models like OpenAI’s GPT and Anthropic’s Claude. Its agents can also be set up to have knowledge of existing product lines or businesses, so they can converse about a client’s offerings.

Video: Fantasy

Fantasy then creates focus groups of both synthetic humans and real people. The participants are given a topic or a product idea to discuss, and Fantasy and its client watch the chatter. BP, an oil and gas company, asked a swarm of 50 of Fantasy’s synthetic humans to discuss ideas for smart city projects. “We've gotten a really good trove of ideas,” says Roger Rohatgi, BP’s global head of design. “Whereas a human may get tired of answering questions or not want to answer that many ways, a synthetic human can keep going,” he says.

Peter Smart, chief experience officer at Fantasy, says that synthetic humans have produced novel ideas for clients, and prompted real humans included in their conversations to be more creative. “It is fascinating to see novelty—genuine novelty—come out of both sides of that equation—it’s incredibly interesting,” he says.

Video: Fantasy

Large language models are proving remarkably good at mirroring human behavior. Their algorithms are trained on huge amounts of text slurped from books, articles, websites like Reddit, and other sources—giving them the ability to mimic many kinds of social interaction.

When these bots adopt human personas, things can get weird.

Most Popular

Experts warn that anthropomorphizing AI is both potentially powerful and problematic, but that hasn’t stopped companies from trying it. Character.AI, for instance, lets users build chatbots that assume the personalities of real or imaginary individuals. The company has reportedly sought funding that would value it at around $5 billion.

The way language models seem to reflect human behavior has also caught the eye of some academics. Economist John Horton of MIT, for instance, sees potential in using these simulated humans—which he dubs Homo silicus—to simulate market behavior.

You don’t have to be an MIT professor or a multinational company to get a collection of chatbots talking amongst themselves. For the past few days, WIRED has been running a simulated society of 25 AI agents go about their daily lives in Smallville, a village with amenities including a college, stores, and a park. The characters’ chat with one another and move around a map that looks a lot like the game Stardew Valley. The characters in the WIRED sim include Jennifer Moore, a 68-year-old watercolor painter who putters around the house most days; Mei Lin, a professor who can often be found helping her kids with their homework; and Tom Moreno, a cantankerous shopkeeper.

The characters in this simulated world are powered by OpenAI’s GPT-4 language model, but the software needed to create and maintain them was open sourced by a team at Stanford University. The research shows how language models can be used to produce some fascinating and realistic, if rather simplistic, social behavior. It was fun to see them start talking to customers, taking naps, and in one case decide to start a podcast.

Large language models “have learned a heck of a lot about human behavior” from their copious training data, says Michael Bernstein, an associate professor at Stanford University who led the development of Smallville. He hopes that language-model-powered agents will be able to autonomously test software that taps into social connections before real humans use them. He says there has also been plenty of interest in the project from videogame developers, too.

Video: Fantasy

The Stanford software includes a way for the chatbot-powered characters to remember their personalities, what they have been up to, and to reflect upon what to do next. “We started building a reflection architecture where, at regular intervals, the agents would sort of draw up some of their more important memories, and ask themselves questions about them,” Bernstein says. “You do this a bunch of times and you kind of build up this tree of higher-and-higher-level reflections.”

Anyone hoping to use AI to model real humans, Bernstein says, should remember to question how faithfully language models actually mirror real behavior. Characters generated this way are not as complex or intelligent as real people and may tend to be more stereotypical and less varied than information sampled from real populations. How to make the models reflect reality more faithfully is “still an open research question,” he says.

Smallville is still fascinating and charming to observe. In one instance, described in the researchers’ paper on the project, the experimenters informed one character that it should throw a Valentine’s Day party. The team then watched as the agents autonomously spread invitations, asked each other out on dates to the party, and planned to show up together at the right time.

WIRED was sadly unable to re-create this delightful phenomenon with its own minions, but they managed to keep busy anyway. Be warned, however, running an instance of Smallville eats up API credits for access to OpenAI's GPT-4 at an alarming rate. Bernstein says running the sim for a day or more costs upwards of a thousand dollars. Just like real humans, it seems, synthetic ones don’t work for free.

You Might Also Like …

Will Knight is a senior writer for WIRED, covering artificial intelligence. He writes the Fast Forward newsletter that explores how advances in AI and other emerging technology are set to change our lives—sign up here. He was previously a senior editor at MIT Technology Review, where he wrote about fundamental… Read more
Senior Writer

More from WIRED

The Meta AI Chatbot Is Mark Zuckerberg's Answer to ChatGPT

Meta's AI assistant can do things like suggest travel plans in a group chat. The company also announced a string of chatbots modeled on celebrities like Snoop Dogg and Paris Hilton.

Enough Talk, ChatGPT—My New Chatbot Friend Can Get Things Done

An experimental AI assistant called Auto-GPT can use the web to solve problems. When the automated helper works, it can feel like the future.

Chatbot Hallucinations Are Poisoning Web Search

Untruths spouted by chatbots ended up on the web—and Microsoft's Bing search engine served them up as facts. Generative AI could make search harder to trust.

Get Ready for AI Chatbots That Do Your Boring Chores

Move over, Siri. Startups are using the technology behind ChatGPT to build more capable AI agents that can control your computer and access the web to get things done—with sometimes chaotic results.

Amazon Upgrades Alexa for the ChatGPT Era

A sweeping upgrade to Amazon’s Alexa taps AI technology like that behind ChatGPT and also allows the virtual assistant to attempt to read body language.

AI Chatbots Can Guess Your Personal Information From What You Type

The AI models behind chatbots like ChatGPT can accurately guess a user’s personal information from innocuous chats. Researchers say the troubling ability could be used by scammers or to target ads.

How ChatGPT Can Help You Do More With PDFs

The AI chatbot can search, summarize, and create PDF documents with a few handy plug-ins.

OpenAI’s Dall-E 3 Is an Art Generator Powered by ChatGPT

OpenAI has launched Dall-E 3, which uses ChatGPT to take the pain out of prompting. Now you can modify artwork by simply talking to the chatbot.

*****
Credit belongs to : www.wired.com

Check Also

Dubai deluge likely made worse by warming world, scientists find

A powerful rainstorm that wreaked havoc on the desert nation of the United Arab Emirates …