AIResearch AIResearch
Back to articles
Coding

AI Learns When to Think Deeply or Briefly

New method helps artificial intelligence solve complex problems more efficiently by automatically adjusting its reasoning approach based on difficulty

AI Research
November 05, 2025
3 min read
AI Learns When to Think Deeply or Briefly

Large reasoning models, the advanced AI systems capable of complex problem-solving, often struggle with cognitive inefficiencies—either overthinking simple questions or underthinking difficult ones. A new approach called DeepCompress addresses this fundamental limitation by teaching AI to dynamically adjust its reasoning process, achieving both better accuracy and faster performance across mathematical benchmarks.

Researchers discovered that AI models can solve problems more effectively when they match their reasoning approach to the difficulty level. The DeepCompress method enables models to automatically classify problems as either "simple" or "hard" in real-time, then adjust their reasoning strategy accordingly. For problems the model finds easy, it uses shorter, more efficient reasoning chains. For challenging problems, it employs longer, more exploratory reasoning paths that increase the likelihood of finding correct solutions.

The methodology builds on reinforcement learning techniques, using a dual-reward system that encourages different behaviors for different problem types. The system monitors the model's performance on batches of problems, calculating what percentage of responses are correct. When the model performs well on a particular problem type (achieving over 50% accuracy), those problems are classified as "simple" and the model receives rewards for shorter responses. When performance drops below this threshold, problems are labeled "hard" and the model is incentivized to use more extensive reasoning.

Experimental results demonstrate significant improvements across seven mathematical reasoning benchmarks. DeepCompress achieved state-of-the-art performance while reducing average response length by 57.9% for the 3-billion parameter model and 16.6% for the 7-billion parameter model compared to previous methods. On particularly challenging problems from the American Invitational Mathematics Examination, the method improved accuracy by 6.5 percentage points while using 35.2% fewer tokens. The analysis revealed that models using DeepCompress showed more frequent reflection behaviors—systematically checking intermediate steps and revising approaches—while maintaining shorter overall responses, indicating more efficient problem-solving.

This advancement matters because it addresses a critical bottleneck in AI deployment: the computational cost of running large reasoning models. By making AI reasoning more efficient without sacrificing accuracy, the method could enable broader applications in education, research, and problem-solving domains where both performance and cost matter. The approach demonstrates that smarter thinking, not just more thinking, leads to better results.

The method does have limitations. Its effectiveness depends on having sufficient variation in problem difficulty within training batches, and the researchers capped generation length at 10,000 tokens, which may restrict exploration of extremely complex solutions requiring longer reasoning chains. Future work could explore how to maintain these efficiency gains while allowing for even more extensive reasoning when truly necessary.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn