MIT Technology Review May 8, 2024
MIT Technology Review Insights

AI models that process multiple types of information at once bring even bigger opportunities, along with more complex challenges, than traditional unimodal AI.

Multimodality is a relatively new term for something extremely old: how people have learned about the world since humanity appeared. Individuals receive information from myriad sources via their senses, including sight, sound, and touch. Human brains combine these different modes of data into a highly nuanced, holistic picture of reality.

“Communication between humans is multimodal,” says Jina AI CEO Han Xiao. “They use text, voice, emotions, expressions, and sometimes photos.” That’s just a few obvious means of sharing information. Given this, he adds, “it is very safe to assume that future communication between human and machine will...

Today's Sponsors

LEK
ZeOmega

Today's Sponsor

LEK

 
Topics: AI (Artificial Intelligence), Technology
OpenAI Bosses: We Take Safety ‘Very Seriously’
Microsoft set to unveil its vision for AI PCs at Build developer conference
Mastering prompt engineering to elevate AI in clinical practice
AI-Guidance Boosts Arrhythmia Ablation Success in Persistent Cases
HealthTech Acquirers Key Priorities Mid 2024

Share This Article