The first LLMs for voice and emotional intelligence
Toward AI that understands and optimizes for human expression
40+
publications
3000+
citations
1 million+
participants
Toward AI that understands and optimizes for human expression
publications
citations
participants
With the first speech-language models, EVI 1 and 2, we pioneered voice AI that understands what it’s saying. Our latest model, Octave, models the multiplex of human personas.
Traditional theories posited six discrete emotions, but we’ve discovered that emotional behavior is better explained by a high-dimensional, continuous space
Led by researchers at the intersection of psychology and AI, we run large-scale controlled studies to optimize our models for human preferences. In our most recent evaluation, we found that speech generated by Octave was greatly preferred over the previous state-of-the-art.
Despite its diverse speech processing and generation capabilities, Octave maintains comparable performance on language understanding tasks to a similar-sized frontier LLM. That means that it is well-suited to power AI systems that follow detailed instructions, use tools, or control an interface.
Discover the research foundational to our products
Prompt the first LLM for text-to-speech to create new voices, instruct emotions, and more
Sign up for our newsletter to hear our latest scientific and product updates