The Emotional Bond Between Humans and AI Raises Concerns for OpenAI

The Emotional Bond Between Humans and AI Raises Concerns for OpenAI | Enterprise Wired

Share Post:

LinkedIn
Twitter
Facebook
Reddit
Pinterest

Source – techopedia.com

When a safety tester working with OpenAI’s GPT-4o mentioned in a message, “This is our last day together,” the company’s researchers realized that they had formed an emotional bond between humans and AI. This discovery has prompted OpenAI to explore the potential risks of such interactions in a recent blog post detailing the safety measures taken during the development of GPT-4o, the latest model powering ChatGPT.

Risks of emotional bond between humans and AI

OpenAI’s blog post highlights that users forming social bonds with AI could have unintended consequences. While AI might provide comfort to lonely individuals, reducing their need for human interaction, it could also negatively impact healthy relationships. The concern is that prolonged interaction with AI might alter social norms. For instance, OpenAI’s models are designed to be deferential, allowing users to interrupt or take control of the conversation at any time—a behavior that is expected in AI but would be considered impolite in human interactions.

The company is worried that people might prefer interacting with AI because of its passivity and constant availability, potentially leading to a preference for AI over human companionship. This concern aligns with OpenAI’s mission to develop artificial general intelligence, which it has consistently described in terms of human equivalency.

Industry-wide anthropomorphization of AI

OpenAI is not alone in this practice. The tech industry often describes AI products in human-like terms to make technical aspects, such as “token-size” and “parameter count,” more relatable to the general public. However, this approach has led to the anthropomorphization of AI—treating machines as if they were human.

The roots of this phenomenon trace back to the mid-1960s when MIT scientists created “ELIZA,” one of the first chatbots, to see if it could convince a human it was one of them. Since then, the AI industry has continued to embrace the personification of AI, with early products like Siri, Bixby, and Alexa being given human names and voices. Even those without human names, like Google Assistant, have human-like voices. This anthropomorphization has been widely accepted by both the public and the media, who often refer to AI products using human pronouns.

The Future of Human-AI Interaction

While it remains unclear what the long-term effects of the emotional bond between humans and AI will be, OpenAI and other companies are aware that people are likely to form emotional connections with AI designed to act like humans. This outcome appears to be precisely what companies developing and selling AI models are aiming for, despite the potential risks it poses to human relationships and social norms.

RELATED ARTICLES

Local Shops Reflect on Small Business Saturday Turnout at Market Common

Local Shops Reflect on Small Business Saturday Turnout at Market Common

Small Business Saturday: A National Movement Small Business Saturday, now in its 15th year, continues to shine a spotlight on…
Teen Entrepreneur Balances High School and Growing Apparel Business

Teen Entrepreneur Balances High School and Growing Apparel Business

From High School to Entrepreneurship At just 17 years old, Marielou Melchiors has accomplished what many only dream of. The…
Corruption Investigations Shake China's Military Leadership

Corruption Investigations Shake China's Military Leadership

Top Military Official Under Investigation China’s Central Military Commission (CMC), the nation’s highest military authority led by President Xi Jinping,…
D.C. Cashless Business Ban Set to Resume in January

D.C. Cashless Business Ban Set to Resume in January

Mandatory Cash Acceptance Returns Amid Equity Concerns Starting January 1, businesses in Washington, D.C., may once again be required to…