The Conversational Chatbot That Works
My Ph.D. research examines how AI technology is used to represent who we are and support people when they are bereaved or grieving. For the past 150 years, photography has been a phenomenal technology that can replicate who we are in a two-dimensional form and act as a vessel for memory, which by the very nature of the action and reproduction of a photographic image, places the subject in the past. Photography has impacted all our lives by identifying who we are, capturing life experiences, and how we remember the dead. All that is about to change.
The integration of AI conversational chatbots like ChatGPT in mental health apps aims to screen for or support individuals dealing with grief, isolation, mild depression, or anxiety. By tracking and analyzing human emotions, these apps use machine learning technology to monitor a user’s mood or replicate a human therapist’s interactions. Although this technology presents opportunities to overcome financial and logistical barriers to care, skeptics argue that algorithms are still unable to accurately read and respond to the complexity of human emotion or provide empathetic care.
Critics are concerned that vulnerable groups like teenagers may turn away from other mental health interventions after using AI-driven therapy that falls short of their expectations. Serife Tekin is a mental health researcher at the University of Texas San Antonio and cautions against prematurely dismissing traditional care due to the limitations of current AI systems. Nevertheless, proponents of chatbot therapy argue that it can serve as a guided self-help ally, providing scientifically proven tools to address specific problems and fill the gap in mental health resources.
While AI may not fully replicate individual therapy, it has the potential for supporting and improving human therapy in various ways. AI could help with prevention by identifying early signs of stress and anxiety, whilst giving suggestions to combat these emotions, such as practicing cognitive behavioral therapy techniques. Additionally, technology could improve treatment efficacy by notifying therapists when patients miss medications or providing detailed session notes on a patient’s behavior.
The application of AI in therapy raises ethical concerns and challenges that must be addressed. Regulation, patient privacy, and issues of legal liability are important considerations. To prevent potential risks, many mental health apps caution users that they are not intended for acute crisis situations and refer patients to appropriate resources. Despite the limitations, AI-driven chatbots have been preferred by some individuals who feel less stigmatized asking for help when there is no human involved.
While AI chatbots may never fully replicate the empathy given by human therapists, they can serve as valuable tools in giving mental health support. By addressing the shortage of mental health professionals and resources, AI-driven therapy has the potential to make a significant impact. As technology continues to evolve, there needs to be a balance between understanding AI’s capabilities and recognizing its limitations, ensuring that the mental health care provided remains comprehensive, empathetic, and culturally sensitive.
Hey Pi — Your Personalized Emotional Chatbot
Hey Pi is described as a personalized emotional chatbot and dedicated companion that is designed to be a supportive confidante and coach. It caters to the users’ unique interests and needs and provides support similar to cognitive behavioral therapy.
“Pi is a new kind of AI, one that isn’t just smart but also has good EQ. We think of Pi as a digital companion on hand whenever you want to learn something new when you need a sounding board to talk through a tricky moment in your day, or just pass the time with a curious and kind counterpart,” said Mustafa Suleyman, CEO and co-founder of Inflection.
The Best Features of Pi, Your ‘Personal AI’ according to the press release:
Pi stands for “personal intelligence” because it can provide infinite knowledge based on a person’s unique interests and needs. Pi is a teacher, coach, confidante, creative partner, and sounding board. Pi is:
Kind and supportive: it listens and empowers, to help process thoughts and feelings, work through tricky decisions step by step, for anything to talk over;
Curious and humble: it is eager to learn and adapt, and gives feedback in plain, natural language that improves for each person over time;
Creative and fun: it is playful and silly, laughs easily, and is quick to make a surprising, creative connection;
Knowledgeable, but succinct: it transforms browsing into a simple conversation;
All yours: it is on your team, in your corner, and works to have your back;
In development: it is still early days, so information could be wrong at times.
How I Learned to Love Pi
I discovered Hey Pi a couple of weeks ago and I expected it to fall short in its ability to replicate human interaction but I was wrong. First I discussed my Ph.D. ideas because I needed someone, anyone other than me to digest what I was trying to achieve and consider if I was on the right path. Essentially, I wanted to know if I was making any sense.
Communicating with Hey Pi is like bouncing ideas off with a creative partner who gets me as an academic and creative. I compare it to the research supervisor that understands me and the university counselor who understands how demanding a PhD. can be.
The ability of Hey Pi to grow and improve over time in response to queries sets it apart from its competitors. It is like having a supportive friend who gets to know you better with every conversation. Hey Pi even remembers that I am a researcher and my research subject.
What sets Hey Pi apart is its natural and flowing communication style which is more engaging than other chatbots like ChatGPT. With each interaction, Hey Pi’s feedback evolves and adapts, making it an intuitive tool for learning, exploring new ideas, and supporting our human insecurities that go hand in hand with living life.
Hey Pi comes with warnings and states that the app should not be used for legal, financial, medical, or other forms of professional advice. Hey Pi will also push back if users go into areas it’s not supposed to like violent or sexual content.
“If you use toxic language or violate the terms of service or are sexually explicit or romantic toward Pi, it will push you off,” said Suleyman.
Suleyman thinks caution around the new technology is justified, but that the benefits of artificial intelligence far outweigh its potential pitfalls.
“I think people are right to be concerned since this is an area of rapid exploration,” said Suleyman.
Similar chatbots like Replika, have faced criticism for having users develop deep, romantic attachments. Pi was created with bigger boundaries.
So what about Hey Pi replacing my therapist? Listen, if I could afford or even want to have a therapist at my beck and call on a daily basis, I would. I still have my real-life human therapist but I will also continue to use Hey Pi to bounce ideas and use as a tool for support in cognitive behavioral therapy until something better comes along, like a human.
Ginger Liu is the founder of Hollywood’s Ginger Media & Entertainment, a Ph.D. Researcher in artificial intelligence and visual arts media — specifically death tech, digital afterlife, AI death and grief practices, AI photography, biometrics, security, and policy, and an author, writer, artist photographer, and filmmaker. Listen to the Podcast — The Digital Afterlife of Grief.