Talk

How to crack 500ms latency in AI voicebots?

San Francisco

500ms is the magic number for fluid voice interactions, but most voicebots fall short, leading to frustrating lag. We've solved the latency puzzle by using every optimization possible to combine latest LLMs, speech-to-text, and text-to-speech while seamlessly handling interruptions and backchanneling as appropriate.

We'll share our journey from 2500ms to sub-500ms latency, detailing the key techniques that enabled this 5x speedup. From efficient model architectures to real-time audio streaming to contextual cues, you'll learn proven optimizations for building responsive, human-like voicebots. Join us to explore the cutting edge of low-latency conversational AI and take away practical insights for your own projects.

Speaker

Video