PYMNTS.com October 18, 2024

Recent findings from Apple researchers have cast doubt on the mathematical prowess of large language models (LLMs), challenging the notion that artificial intelligence (AI) is on the brink of human-like reasoning.

In a test of 20 state-of-the-art LLMs, performance on grade-school math problems plummeted when questions were slightly modified or irrelevant information was added, Apple found. Accuracy dropped by up to 65.7%, revealing a startling fragility in AI systems when faced with tasks requiring robust logical reasoning.

This weakness could have far-reaching implications for commerce relying on AI for complex decision-making. Financial institutions, in particular, may need to reassess their use of AI in tasks involving intricate calculations or risk assessment.

At the heart of this debate lies the artificial...

Today's Sponsors

LEK
ZeOmega

Today's Sponsor

LEK

 
Topics: AI (Artificial Intelligence), Technology
Google digs deeper into healthcare AI: 5 notes
JP Morgan Annual Healthcare Conference 2025: What are the key talking points likely to be?
How AI Has And Will Continue To Transform Healthcare
AI Translates Nature into New Medicines | StartUp Health Insights: Week of Nov 26, 2024
Building AI trust: The key role of explainability

Share This Article