AI Systems Struggle with Basic Time-Telling Tasks, Study Reveals

Curated by THEOUTPOST

On Fri, 14 Mar, 8:05 AM UTC

6 Sources

Share

A study by University of Edinburgh researchers shows that advanced AI models have difficulty interpreting analog clocks and calendars, highlighting a significant gap in AI capabilities for everyday tasks.

AI Systems Struggle with Basic Time-Telling Tasks

A recent study conducted by researchers at the University of Edinburgh has revealed a surprising limitation in advanced artificial intelligence (AI) systems: they struggle to perform basic time-telling tasks that most humans learn at an early age. The study, led by Rohit Saxena from the School of Informatics, tested various state-of-the-art AI models on their ability to interpret analog clocks and calendars 1.

Clock Reading Challenges

The research team evaluated several multimodal large language models (MLLMs), including systems from Google DeepMind, Anthropic, Meta, Alibaba, ModelBest, and OpenAI. These AI models were presented with images of different clock designs, including those with Roman numerals, varying dial colors, and with or without second hands 2.

The results were striking:

  • AI systems correctly interpreted clock-hand positions less than 25% of the time
  • Performance worsened with Roman numerals and stylized clock hands
  • Removing the second hand did not improve accuracy, suggesting fundamental issues with hand detection and angle interpretation 3

Calendar Comprehension Difficulties

The study also tested the AI models' ability to answer calendar-based questions, such as identifying holidays and calculating dates. Even the best-performing AI model made errors in date calculations 20% of the time 4.

Implications for AI Development

This research highlights a significant gap between AI's capabilities in complex tasks and its struggles with everyday skills that humans often take for granted. Aryo Gema, another researcher involved in the study, noted:

"AI research today often emphasizes complex reasoning tasks, but ironically, many systems still struggle when it comes to simpler, everyday tasks. Our findings suggest it's high time we addressed these fundamental gaps." 5

Future Applications and Challenges

The ability to interpret time from visual inputs is crucial for many real-world applications, including:

  • Scheduling assistants
  • Autonomous robots
  • Tools for people with visual impairments

Overcoming these limitations could significantly enhance AI's integration into time-sensitive, real-world applications. However, the current shortfalls present a notable obstacle to achieving this goal 1.

The findings of this study will be presented at the Reasoning and Planning for Large Language Models workshop at The Thirteenth International Conference on Learning Representations (ICLR) in Singapore on April 28, 2025, highlighting the importance of addressing these fundamental gaps in AI capabilities.

Continue Reading
Apple Study Reveals Limitations in AI's Mathematical

Apple Study Reveals Limitations in AI's Mathematical Reasoning Abilities

A recent study by Apple researchers exposes significant flaws in the mathematical reasoning capabilities of large language models (LLMs), challenging the notion of AI's advanced reasoning skills and raising questions about their real-world applications.

PYMNTS.com logoWired logoFuturism logoTechRadar logo

17 Sources

PYMNTS.com logoWired logoFuturism logoTechRadar logo

17 Sources

Apple Research Exposes Fundamental Flaws in AI's Logical

Apple Research Exposes Fundamental Flaws in AI's Logical Reasoning Capabilities

Apple researchers conducted tests revealing significant limitations in AI models' ability to perform simple arithmetic and logical reasoning, raising questions about the true intelligence of current AI systems.

9to5Mac logoMiami Herald logo

2 Sources

9to5Mac logoMiami Herald logo

2 Sources

New AGI Benchmark Stumps Leading AI Models, Highlighting

New AGI Benchmark Stumps Leading AI Models, Highlighting Gap in General Intelligence

The Arc Prize Foundation introduces ARC-AGI-2, a challenging new test for artificial general intelligence that current AI models, including those from OpenAI and Google, are struggling to solve. The benchmark emphasizes efficiency and adaptability, revealing limitations in current AI capabilities.

TechCrunch logoNew Scientist logoTom's Guide logoMashable logo

5 Sources

TechCrunch logoNew Scientist logoTom's Guide logoMashable logo

5 Sources

AI Models Struggle with Abstract Visual Reasoning, Falling

AI Models Struggle with Abstract Visual Reasoning, Falling Short of Human Capabilities

A study by USC researchers reveals that AI models, particularly open-source ones, struggle with abstract visual reasoning tasks similar to human IQ tests. While closed-source models like GPT-4V perform better, they still fall short of human cognitive abilities.

ZDNet logoTech Xplore logoScienceDaily logoNeuroscience News logo

4 Sources

ZDNet logoTech Xplore logoScienceDaily logoNeuroscience News logo

4 Sources

AI Falls Short in Understanding Human Social Interactions,

AI Falls Short in Understanding Human Social Interactions, Study Reveals

A new study from Johns Hopkins University shows that current AI models struggle to interpret social dynamics and context in video clips, highlighting a significant gap between human and machine perception of social interactions.

Popular Science logoScienceDaily logoNeuroscience News logoObserver logo

4 Sources

Popular Science logoScienceDaily logoNeuroscience News logoObserver logo

4 Sources

TheOutpost.ai

Your one-stop AI hub

The Outpost is a comprehensive collection of curated artificial intelligence software tools that cater to the needs of small business owners, bloggers, artists, musicians, entrepreneurs, marketers, writers, and researchers.

© 2025 TheOutpost.AI All rights reserved