How Dot uses Hume's API for emotionally intelligent AI
By Lydia Schooler on Aug 22, 2024
Key impact metrics
- Achieved 60% user adoption of voice messages
- Drove 45% conversion rate to paid subscriptions
- Built trust through emotionally intelligent interactions
- Integrated Hume API in less than one day
Nuances of speaking to an AI with voice
Dot, a personal AI developed by New Computer, is designed to grow alongside its users, serving as a thought partner, friend, and confidant for young adults navigating uncertainty in their lives. Built on an infinite long-term memory system, Dot's responses become increasingly personalized with each interaction, whether through speaking or typing.
After introducing a voice message feature, New Computer noticed that Dot interpreted these messages too literally, relying solely on transcriptions. According to Luca Beetz, Founding AI Engineer, this led to “flat, robotic conversations where Dot often misunderstood the user’s intent.” For example, Dot would respond with the same standard helpfulness to a question posed in a defeated tone as it would to one posed in an inquisitive tone.
To truly understand users, Dot needed to be able to pick up on nonverbal cues, so the New Computer team turned to Hume for a scientifically-grounded technical solution.
Finding a solution in Hume's emotionally intelligent AI
To recognize users’ moods, New Computer integrated Hume’s Expression Measurement API. Hume’s expression measurement models – which are built on a decade of research, millions of proprietary data points, and over 30 publications in leading journals – allow Dot to interpret and respond with emotional intelligence to 25 distinct emotions by analyzing vocal tones.
New Computer found integrating the API “extremely easy,” taking “less than a day to integrate and get up and running.”
The impact of empathic AI
-
Hume’s Expression Measurement APl allows Dot to maintain the emotional context of conversations in its long-term memory, enabling it to build a ‘theory of mind’ about its users and how they respond to different topics or under various circumstances.
-
Dot can respond to messages with greater emotional intelligence, making interactions feel more human and empathetic. This builds trust and paves the way for a real relationship to be built, one that moves beyond the transactional messaging interaction that exists between most users and their AI tools today.
-
By detecting and responding to emotional expressions, Dot boosts user engagement and satisfaction, making the app a trusted companion.
Since integrating Hume's Expression Measurement API, over 60% of users have engaged with Dot using voice messages.
“We’ve seen that users want to be able to interact with Dot over voice, like they would with a friend,” says Samantha Whitmore, CEO of New Computer. “Thanks to the emotional interpretations we get from Hume, users are consistently amazed by Dot’s ability to intuitively sense and respond to the emotional undertones in their messages.”
Creating a sense of true, empathetic understanding pays off: over 45% of users that chat with Dot consistently convert to paid subscriptions. New Computer attributes much of this high conversion rate to the sense of being seen and heard that Dot provides through its emotionally intelligent interactions.
Learn More: Start measuring vocal and facial expressions with unmatched precision using Hume’s Expression Measurement API. Instantly capture nuanced expressions in audio, video, and images, such as awkward laughter, sighs of relief, and nostalgic glances. Enhance your product decisions and user experience with models based on 10+ years of research. Start building today.
Subscribe
Sign up now to get notified of any updates or new articles.
Recent articles
Speech-to-text and text-to-speech
Speech-to-text (STT) and text-to-speech (TTS) are two groundbreaking technologies that have transformed how we engage with computers and other devices. Leading tech companies like Google, IBM, and Amazon are constantly competing to develop the most accurate and sophisticated speech recognition systems. While both STT and TTS involve converting between spoken and written language, they have distinct functions and applications. This article explores the inner workings of each technology, examines their diverse use cases, analyzes their strengths and weaknesses, and discusses the current advancements and future trends in the field.
Creating AI voiceovers with emotion
The use of artificial intelligence (AI) in voice generation has progressed significantly, moving beyond the monotonous, robotic voices of the past to more natural and expressive speech. This evolution has opened up exciting possibilities for various applications, including voiceovers. This article delves into current techniques for creating AI voiceovers with emotion, explores the upcoming capabilities of next-generation AI voice models like Hume AI's OCTAVE, and discusses the potential benefits and challenges of this technology.
Creating custom character voices with AI
The use of AI in voice generation is rapidly changing how we interact with technology and consume content. From video games and animated films to audiobooks and virtual assistants, AI-generated voices are becoming increasingly prevalent. One of the most exciting applications of this technology is in the creation of custom character voices. This allows developers and creators to imbue their characters with unique and engaging personalities, enhancing the overall user experience. This article delves into the best ways to create custom character voices using AI, exploring the various technologies, platforms, and customization options available.