VentureBeat October 22, 2020
Kyle Wiggers

Amazon’s Alexa is getting better at recognizing who’s speaking and what they’re speaking about, understanding words through on-device techniques, and leveraging models trained without needing human review. That’s according to automatic speech recognition head Shehzad Mevawalla, who spoke with VentureBeat ahead of a keynote address at this year’s Interspeech conference.

Alexa is now running “full-capability” speech recognition on-device, after previously relying on models many gigabytes in size that required huge amounts of memory and ran on servers in the cloud. That change is because of a move to end-to-end models, Mevawalla said, or AI models that take acoustic speech signals as input and directly output transcribed speech. Alexa’s previous speech recognizers had specialized components that processed inputs in sequence, such...

Today's Sponsors

LEK
ZeOmega

Today's Sponsor

LEK

 
Topics: AI (Artificial Intelligence), Technology, Voice Assistant
Mobile Apps Are Back: Why Voice Technology Will Change Everything In 2025
Technology that listens: Benefits and challenges of voice assistants for the elderly
Implementing and Analyzing the Advantages of Voice AI as Measurement-Based Care (MBC) to Address Behavioral Health Treatment Disparities among Youth in Economically Disadvantaged Communities
Amazon Picks Anthropic to Power Alexa 2.0
AI voice assistant can spot complications in heart patients

Share This Article