View in browser | Your newsletter preferences |
By Will Knight | 10.12.23 |
Hello, authentic humans, This week, let's look at how artificial intelligence is being used to mimic real social interactions, and why some companies believe it may help them invent new things. I also get to play god to a small village of AI characters.
|
'Synthetic Humans' Just Want to Talk 🤖 👪 🔮 |
Lena Anderson isn't a soccer fan, but she does spend a lot of time ferrying her kids between soccer practices and competitive games. "I may not pull out a foam finger and painted face, but soccer does have a place in my life," says the affirmed soccer mom—who also happens to be completely made up. Anderson is a fictional personality played by artificial intelligence software like that powering ChatGPT. Anderson doesn't let her imaginary status get in the way of her opinions, though, and comes complete with a detailed backstory. In a wide-ranging conversation with a human interlocutor, the bot says that it has a 7-year-old son who is a fan of the New England Revolution and loves going to home games at Gillette Stadium in Massachusetts. Anderson claims to think the sport is a wonderful way for kids to stay active and make new friends. |
In another conversation, two more AI characters, Jason Smith and Ashley Thompson, talk to one another about ways that Major League Soccer might reach new audiences. Smith suggests a mobile app with an augmented reality feature showing different views of games. Thompson adds that the app could include "gamification" that lets players earn points as they watch. The three bots are among scores of AI characters that have been developed by Fantasy, a New York company that helps businesses such as LG, Ford, Spotify, and Google dream up and test new product ideas. Fantasy calls its bots Synthetic Humans and says they can help clients learn about audiences, think through product concepts, and even generate new ideas, like the soccer app. "The technology is truly incredible," says Cole Sletten, VP of digital experience at the MLS. "We're already seeing huge value and this is just the beginning." Fantasy uses the kind of machine learning technology that powers chatbots like OpenAI's ChatGPT and Google's Bard to create its synthetic humans. The company gives each agent dozens of characteristics drawn from ethnographic research on real people, feeding them into commercial large Language Models like OpenAI's GPT and Anthropic's Claude. Its agents can also be set up to have knowledge of existing product lines or businesses, so they can converse about a client's offerings.
|
Fantasy then creates focus groups of both synthetic humans and real people. The participants are given a topic or a product idea to discuss, and Fantasy and its client watch the chatter. BP, an oil and gas company, asked a swarm of 50 of Fantasy's synthetic humans to discuss ideas for smart city projects. "We've gotten a really good trove of ideas," says Roger Rohatgi, BP's global head of design. "Whereas a human may get tired of answering questions or not want to answer that many ways, a synthetic human can keep going," he says. Peter Smart, chief experience officer at Fantasy, says that synthetic humans have produced novel ideas for clients, and prompted real humans included in their conversations to be more creative. "It is fascinating to see novelty—genuine novelty—come out of both sides of that equation—it's incredibly interesting," he says. Large language models are proving remarkably good at mirroring human behavior. Their algorithms are trained on huge amounts of text slurped from books, articles, websites like Reddit, and other sources—giving them the ability to mimic many kinds of social interaction. When these bots adopt human personas, things can get weird. |
Experts warn that anthropomorphizing AI is both potentially powerful and problematic, but that hasn't stopped companies from trying it. Character.AI, for instance, lets users build chatbots that assume the personalities of real or imaginary individuals. The company has reportedly sought funding that would value it at around $5 billion. The way language models seem to reflect human behavior has also caught the eye of some academics. Economist John Horton of MIT, for instance, sees potential in using these simulated humans—which he dubs Homo silicus—to simulate market behavior. You don't have to be an MIT professor or a multinational company to get a collection of chatbots talking amongst themselves. For the past few days, I have been watching 25 AI agents go about their daily lives in Smallville, a village with amenities including a college, stores, and a park. The characters' chat with one another and move around a map that looks a lot like the game Stardew Valley. The characters in my sim include Jennifer Moore, a 68-year-old watercolor painter who putters around the house most days; Mei Lin, a professor who can often be found helping her kids with their homework; and Tom Moreno, a cantankerous shopkeeper.
|
The characters in this simulated world are powered by OpenAI's GPT-4 language model, but the software needed to create and maintain them was open sourced by a team at Stanford University. The research shows how language models can be used to produce some fascinating and realistic, if rather simplistic, social behavior. It was fun to see them start talking to customers, taking naps, and in one case decide to start a podcast. Large language models "have learned a heck of a lot about human behavior" from their copious training data, says Michael Bernstein, an associate professor at Stanford University who led the development of Smallville. He hopes that language-model-powered agents will be able to autonomously test software that taps into social connections before real humans use them. He says there has also been plenty of interest in the project from videogame developers, too. The Stanford software includes a way for the chatbot-powered characters to remember their personalities, what they have been up to, and to reflect upon what to do next. "We started building a reflection architecture where, at regular intervals, the agents would sort of draw up some of their more important memories, and ask themselves questions about them," Bernstein says. "You do this a bunch of times and you kind of build up this tree of higher-and-higher-level reflections." |
Anyone hoping to use AI to model real humans, Bernstein says, should remember to question how faithfully language models actually mirror real behavior. Characters generated this way are not as complex or intelligent as real people and may tend to be more stereotypical and less varied than information sampled from real populations. How to make the models reflect reality more faithfully is "still an open research question," he says. Smallville is still fascinating and charming to observe. In one instance, described in the researchers' paper on Smallville, the experimenters informed one character that it should throw a Valentine's Day party. The team then watched as the agents autonomously spread invitations, asked each other out on dates to the party, and planned to show up together at the right time. I wasn't able to re-create this delightful phenomenon with my minions, sadly, but it was certainly fun to watch them milling around. Be warned, however, running an instance of Smallville eats up API credits for access to OpenAI's GPT-4 at an alarming rate. Bernstein says running sim for a day or more costs upwards of a thousand dollars. Just like real humans, it seems, synthetic ones don't work for free. |
|
|
| Will Knight, Senior Writer (@willknight) |
|
|
Generative AI Is Coming for Sales Execs' Jobs—and They're Celebrating |
ChatGPT-style AI can tackle the drudge work of responding to RFPs faster than humans. Sales teams at Google, Twilio, and others say productivity is spiking. |
| New York's Airbnb Ban Is Descending Into Pure Chaos |
People are listing short-term rentals on social media and lesser-known platforms, bolstering a rental black market in New York City. |
|
|
In the War Against Russia, Some Ukrainians Carry AK-47s. Andrey Liscovich Carries a Shopping List |
Kyiv enlisted a Silicon Valley insider to rush consumer-grade tech onto the battlefield. He's giving a demo of the future of war: the military-retail complex. |
| Inside FTX's All-Night Race to Stop a $1 Billion Crypto Heist |
The same chaotic day FTX declared bankruptcy, someone began stealing hundreds of millions of dollars from its coffers. A WIRED investigation reveals the company's "very crazy night" trying to stop them. |
|
|
For all our future-gazing tech coverage, visit WIRED Business. |
|
|
|