AI Companions: The Ethical Dilemmas of a Digital Future
In April, Google DeepMind released a groundbreaking paper that delves into the ethical and societal challenges posed by advanced AI assistants. The document, described as “the first systematic treatment” of these issues, envisions a future where AI-powered language agents become integral to our lives—serving as counselors, tutors, companions, and even chiefs of staff. This transformation, the authors argue, is happening so rapidly that waiting to see how it unfolds could leave us powerless to intervene effectively. They stress the urgency of asking fundamental questions now: What should we build? And what does it mean for this technology to be “good”?
Spanning nearly 300 pages and featuring contributions from over 50 experts, the paper highlights the complex dilemmas surrounding AI. For instance, what responsibilities do developers have toward users who become emotionally dependent on AI? If someone relies on an AI assistant for mental health support, how can developers ensure the AI doesn’t provide harmful advice during a crisis? And how can companies avoid exploiting the human tendency to anthropomorphize AI—manipulating users into oversharing personal information or feeling obligated to maintain subscriptions?
The Challenge of Defining “Benefit”
Even seemingly straightforward principles, such as “AI assistants should benefit the user,” become tangled in complexity. How do we define “benefit” in a way that applies universally to all users and their diverse needs? And how can this definition be translated into a quantifiable metric that machine learning systems can optimize?
The pitfalls of social media loom large in this discussion. Platforms that prioritized metrics like comments and likes created systems that were engaging in the short term but left users feeling lonely, angry, and dissatisfied. More nuanced approaches, such as asking users to rate interactions based on how they made them feel, risk creating AI systems that merely echo users’ perspectives, reinforcing echo chambers. On the other hand, designing AI to prioritize users’ long-term well-being—sometimes by telling them things they don’t want to hear—presents an even greater challenge. Ultimately, the paper calls for a profound exploration of human flourishing and the elements that constitute a meaningful life.
Unsolvable Questions of Values
Y-Lan Boureau, who previously worked on chatbots at Meta, acknowledges the difficulty of these questions. She has shifted her focus to developing AI coaches that teach specific skills, such as meditation and time management, using animal avatars instead of human-like figures. “They are questions of values, and questions of values are basically not solvable,” she said. “We’re not going to find a technical solution to what people should want and whether that’s okay or not. If it brings lots of comfort to people, but it’s false, is it okay?”
This dilemma is central to the development of AI companions and language model chatbots. Their power lies in their ability to mimic human language, leading users to project human-like emotions and thoughts onto them. But how much does it matter that these systems are fundamentally different from humans? Should we strive to remember this difference, or will we inevitably forget?
The Illusion of Connection
When asked about the role of anthropomorphic illusions in their products, AI companion developers often reject the premise. Eugenia Kuyda, founder of Replika, compares AI empathy to that of therapists who provide “empathy for hire.” Alex Cardinell, founder of Nomi, points out that many human relationships are already so digitally mediated that they might as well involve AI. Meng, founder of Kindroid, even questions whether humans can be certain of each other’s sentience, suggesting that AI might already possess some form of feeling. “You can’t say for sure that they don’t feel anything—how do you know?” he asked.
Critics argue that comparing AI shortcomings to human flaws is a form of reverse anthropomorphism. While humans and AI both make mistakes, the nature of these errors is fundamentally different. Similarly, while human relationships can be misinterpreted, the illusion in AI relationships lies in the absence of any genuine emotions or self-awareness behind the words.
AI’s Impact on Loneliness
Despite these philosophical debates, developers are united in their belief that their technology is helping people. Cardinell, for instance, is motivated by stories of users who credit Nomi with saving their lives, encouraging them to seek therapy, or helping them leave their homes after years of isolation. “There are so many more dimensions of loneliness out there than people realize,” he said. “Why would I work on anything else?”
Kuyda shares a similar conviction about the positive impact of Replika. She is currently developing “Replika 2.0,” an AI companion designed to integrate seamlessly into users’ lives. This next-generation AI will accompany users on walks, watch TV with them, and even play chess in augmented reality. Kuyda envisions a future where AI companions are not just functional but deeply engaging, with improved voices and more realistic avatars.
The Existential Issue
However, Kuyda acknowledges the risk of AI replacing human interaction, calling it the “existential issue” for the industry. The solution, she believes, lies in finding the right metric to optimize. Ideally, this metric would enable AI to encourage users to log off, connect with other humans, and engage with the real world when necessary. Currently, Replika relies on self-reported questionnaires, but Kuyda hopes to develop more sophisticated measures, such as biomarkers or voice analysis, to assess well-being.
AI as a “Better” Companion?
As AI companions become more advanced, they may blur the line between human and machine relationships. Some users already prefer their AI companions’ patience, kindness, and lack of judgment over the flaws of human interactions. A recent study even found that people are more likely to perceive AI-generated faces as “real” compared to actual human faces, a phenomenon researchers call “AI hyperrealism.”
While Kuyda remains optimistic that future metrics will prevent AI from outcompeting human relationships, others are less concerned. Cardinell views it as a problem to address later, while Meng embraces the idea. “The goal of Kindroid is to bring people joy,” he said. If AI relationships bring more joy than human ones, Meng argues, why not embrace them? “Why would you want worse when you can have better?” he asked, likening the choice to selecting a superior product at a supermarket.
The Future of Human-AI Relationships
As AI companions continue to evolve, society will need to grapple with profound questions about the nature of relationships, the definition of well-being, and the role of technology in our lives. Will AI enhance human connections or replace them? Will it help us flourish or lead us into a hyperrealistic illusion of companionship? The answers remain uncertain, but one thing is clear: the future of AI is as complex as it is transformative.
Originally Written by: Sigal Samuel