The first LLMs for voice and emotional intelligence
Toward AI that understands and optimizes for human expression
40+
publications
3000+
citations
1 million+
participants
Speech recognition, understanding, and generation with the same core intelligence
With the first speech-language models, EVI 1 and 2, we pioneered voice AI that understands what it’s saying. Our latest model, Octave, models the multiplex of human personas.
Fine-tuned with scientifically controlled data
Traditional theories posited six discrete emotions, but we’ve discovered that emotional behavior is better explained by a high-dimensional, continuous space
1/3
Voice AI optimized for human preferences
Led by researchers at the intersection of psychology and AI, we run large-scale controlled studies to optimize our models for human preferences. In our most recent evaluation, we found that speech generated by Octave was greatly preferred over the previous state-of-the-art.

Maintaining frontier language capability
Despite its diverse speech processing and generation capabilities, Octave maintains comparable performance on language understanding tasks to a similar-sized frontier LLM. That means that it is well-suited to power AI systems that follow detailed instructions, use tools, or control an interface.

Publications
Discover the research foundational to our products