PYMNTS.com October 18, 2024
Recent findings from Apple researchers have cast doubt on the mathematical prowess of large language models (LLMs), challenging the notion that artificial intelligence (AI) is on the brink of human-like reasoning.
In a test of 20 state-of-the-art LLMs, performance on grade-school math problems plummeted when questions were slightly modified or irrelevant information was added, Apple found. Accuracy dropped by up to 65.7%, revealing a startling fragility in AI systems when faced with tasks requiring robust logical reasoning.
This weakness could have far-reaching implications for commerce relying on AI for complex decision-making. Financial institutions, in particular, may need to reassess their use of AI in tasks involving intricate calculations or risk assessment.
At the heart of this debate lies the artificial...