Episode 16 Empathy and Digital Health | The Feelings Lab
Published on Mar 8, 2022
This week, we are joined by Dr. Dennis Wall, Professor at Stanford and founder of Cognoa, maker of the first FDA-approved AI-powered diagnosis method for autism, to discuss how empathic AI technologies are poised to transform healthcare.
Join Dr. Wall, Dr. Alan Cowen, CEO of Hume, Dr. Dacher Keltner, Professor of Psychology at UC Berkeley, and Matt Forte as they discuss a future in which conditions like autism can be diagnosed efficiently at home, at a younger (more treatable) age, with the help of empathic AI and crowdsourced empathy. What will healthcare look like when AI-empowered doctors can treat exponentially more patients at a fraction of the current cost? How will AI bring more personalized, accurate diagnosis and treatment to broader, more diverse communities? We provide answers to these questions and more in our episode on Empathy and Digital Health.
First hear Dr. Dennis Wall discusses how AI will transform medical diagnosis. What if we could use AI algorithms that approximate the unique intuitions of medical specialists as a funnel for specialist review, making the diagnostic process more efficient and scalable?
Next hear Dr. Dacher Keltner discusses the potential for data-driven assessments, such as the emotional mimicry methods being developed at Hume AI, to more accurately characterize social and emotional regulation disorders and enable individualized treatments.
Later, Hume AI CEO Dr. Alan Cowen explains the promise of data-driven assessments in every aspect of clinical care, including the value of desegregating patient populations to enable fine-grained dimensional assessments of symptoms and quantifying how problematic each symptom actually is with the help of empathic AI.
Subscribe
Sign up now to get notified of any updates or new articles.
Share article
Recent articles
Introducing Hume’s Empathic Voice Interface (EVI) API
Last month, we released the demo of our Empathic Voice Interface (EVI). The first emotionally intelligent voice AI API is finally here! EVI does a lot more than stitch together transcription, LLMs, and text-to-speech. With a new empathic LLM (eLLM) that processes your tone of voice, EVI unlocks new capabilities like knowing when to speak, generating more empathic language, and intelligently modulating its own tune, rhythm, and timbre. EVI is the first voice AI that really sounds like it understands you.
What is semantic space theory?
Our models and products are built on a cutting-edge approach to understanding emotion: semantic space theory (SST), which uses computational methods and data-driven approaches to map the full spectrum of our feelings