When AI reasoning goes wrong: Microsoft Research shows more tokens can mean more problems
- Microsoft Research investigated the effectiveness of scaling methods for AI reasoning tasks.
- Enterprises aim to integrate AI reasoning, but scaling benefits vary across models and tasks.
- The analysis covered nine models on eight benchmark datasets, including math and planning.
- Researchers found DeepSeek-R1 used five times the tokens of Claude 3.7 Sonnet on AIME 2025.
- The study suggests that more compute does not guarantee better results, thus purposeful scaling is needed.
5 Articles
5 Articles


When AI reasoning goes wrong: Microsoft Research shows more tokens can mean more problems
Not all AI scaling strategies are equal. Longer reasoning chains are not sign of higher intelligence. More compute isn't always the answer.
Research Suggests Powerful AI Models Now Possible Without High-End Hardware
Large language models (LLMs) often require substantial computing resources that are usually met by high-performance hardware. These systems are built to handle vast amounts of data and execute the intricate calculations that power these models. For most people, the prospect of running advanced AI technology on their everyday devices seems unrealistic. However, a recent collaborative effort by researchers from MIT, the King Abdullah University o…
Coverage Details
Bias Distribution
- 67% of the sources are Center
To view factuality data please Upgrade to Premium
Ownership
To view ownership data please Upgrade to Vantage