Ƶ

The latest version of ChatGPT has a feature you’ll fall in love with. And that’s a worry

2024-09-12T08:21:00+10:00

Asian woman with cyborg woman or man. AI and human love concept. Enamored girl and humanoid robot couple. Relationship between robot and woman. Romantic relationships with artificial intelligence
Robert Brooks
Robert Brooks,

Even OpenAI is worried about people developing intimate relationships with the new human-like version of its language model-cum-chatbot. But it should have known the risks.

If you’re a paid subscriber to ChatGPT, you may have noticed the artificial intelligence (AI) large language model has recently started to sound more human when you are having audio interactions with it.

That’s because the company behind the language model-cum-chatbot, OpenAI, is currently running a limited pilot of a new feature known as “advanced voice mode”.

“features more natural, real-time conversations that pick up on and respond with emotion and non-verbal cues”. for all paid ChatGPT subscribers to have access to the advanced voice mode in coming months.

Advanced voice mode sounds strikingly human. There aren’t the awkward gaps we are used to with voice assistants; instead it seems to take breaths like a human would. It is also unfazed by interruption, conveys appropriate emotion cues and seems to infer the user’s emotional state from voice cues.

But at the same time as making ChatGPT seem more human, OpenAI that users might respond to the chatbot as if it were human – by developing an intimate relationship with it.

This is not a hypothetical. For example, a social media influencer named Lisa Li . But why exactly do some people develop intimate relationships with a chatbot?

The evolution of intimacy

Humans have a remarkable capacity for friendship and intimacy. This is an extension of the way primates physically to build alliances that can be called upon in times of strife.

But our ancestors also evolved a remarkable capacity . This drove the evolutionary cycle in which the language centres in our brains became larger and what we did with language became more complex.

More complex language in turn enabled more complex socialising with larger networks of relatives, friends and allies. It also enlarged the social parts of our brains.

Language evolved alongside human social behaviour. The way we draw an acquaintance into friendship or a friend into intimacy is largely through conversation.

revealed that conversational back-and-forth, especially when it involves disclosing personal details, builds the intimate sense our conversation partner is somehow part of us.

So I’m not surprised that attempts to replicate this process of “escalating self-disclosure” results in humans feeling .

And that’s just with text input. When the main sensory experience of conversation – voice – gets involved, the effect is amplified. Even voice-based assistants that don’t sound human, such as Siri and Alexa, still get .

The writing was on the lab chalkboard

If OpenAI were to ask me how to ensure users don’t form social relationships with ChatGPT, I would have a few simple recommendations.

First, don’t give it a voice. Second, don’t make it capable of holding up one end of an apparent conversation. Basically don’t make the product you made.

The product is so powerful precisely because it does such an excellent job of mimicking the traits we use to form social relationships.

OpenAI should have known the risks of creating a human-like chatbot.

The writing was on the laboratory chalkboard since the first chatbots flickered on . Computers have been for at least 30 years. The advanced voice mode of ChatGPT is merely the next impressive increment, not what the tech industry would gushingly call a “game changer”.

That users not only form relationships with chatbots but develop very close personal feelings became clear early last year when found themselves unexpectedly cut off from the most advanced functions of their chatbots.

Replika was less advanced than the new version of ChatGPT. And yet the interactions were of such a quality that users formed surprisingly deep attachments.

The risks are real

Many people, for the kind of company that listens in a non-judgmental way, will get a lot out of this new generation of chatbots. They may feel . These kinds of benefits of technology can never be overlooked.

But the potential dangers of ChatGPT’s advanced voice mode are also very real.

Time spent chatting with any bot is time that can’t be spent interacting with friends and family. And people who spend a lot of are at greatest of displacing relationships with other humans.

As OpenAI identifies, chatting with bots can also contaminate existing relationships people have with other people. They may come to expect their partners or friends to behave like polite, submissive, deferential chatbots.

These bigger are going to become more prominent. On the upside, they may also provide deep insights into how culture works.The Conversation

, Scientia Professor of Evolutionary Ecology; Academic Lead of UNSW's Grand Challenges Program,

This article is republished from under a Creative Commons license. Read the .