AIResearch AIResearch
Back to articles
Coding

AI Research Needs More Scientific Rigor, Experts Say

Workshop reveals growing gap between AI theory and practice, calling for better communication between researchers and engineers to advance the field responsibly.

AI Research
November 14, 2025
3 min read
AI Research Needs More Scientific Rigor, Experts Say

As artificial intelligence systems become increasingly integrated into critical applications like self-driving cars and healthcare, a fundamental question emerges: Are we building these powerful technologies on solid scientific foundations? A recent workshop at NeurIPS 2019 brought together leading researchers to address this concern, revealing significant gaps between theoretical understanding and practical implementation in AI. The discussions highlighted that despite AI's remarkable progress, many fundamental questions about how these systems work remain unanswered, creating potential risks as these technologies become more widespread.

The workshop identified several critical challenges facing the AI community. Deep learning systems, while achieving impressive results, operate as a complex mix of art and science rather than being built on well-understood principles. Researchers found that theoretical developments often raise more questions than they answer, and practical applications frequently outpace scientific understanding. This disconnect becomes particularly problematic when AI systems are deployed in safety-critical domains where reliability and explainability matter.

Methodologically, the workshop emphasized three key pillars of machine learning: data, architecture, and optimization algorithms. While significant effort has focused on developing new architectures and algorithms, the role of data remains the least explored area. Researchers discussed how many AI systems rely on massive amounts of labeled data, which is often difficult to obtain automatically. They proposed several approaches to address this, including creating more realistic simulators, developing semi-automatic annotation tools, and leveraging self-supervised learning methods that don't require extensive human labeling.

Analysis of the workshop discussions revealed recurring themes across different AI subfields. In computer vision, researchers identified the need for systems that can generalize to unknown environments and adapt to new conditions while maintaining robustness. For natural language processing, questions emerged about proper benchmarking and detecting when models overfit to training data. Across all domains, participants noted the challenge of evaluating AI systems when test data may not adequately represent real-world conditions, and the difficulty of ensuring models remain reliable when faced with adversarial examples.

The implications of these findings extend beyond academic research. As AI systems are deployed in critical applications like autonomous vehicles and healthcare, the lack of scientific understanding becomes a practical concern. Workshop participants argued that good performance alone isn't enough—explainability and accountability are increasingly necessary. For example, in self-driving car systems, researchers emphasized the importance of holistic approaches that incorporate application-specific knowledge and safety bounds rather than relying solely on data-driven methods.

Several limitations in current AI research were identified. Theoretical approaches often require restrictive assumptions that may not hold in practical applications, while empirical methods may lack rigorous guarantees. The workshop noted that cultural differences between theoretical and applied researchers sometimes lead to one approach being systematically devalued, hindering progress. Additionally, the rapid pace of AI development creates challenges for maintaining scientific rigor, with participants noting that it's becoming increasingly difficult to keep track of publications and ensure research quality.

The workshop concluded that bridging the gap between AI theory and practice requires sustained effort across multiple fronts. Researchers called for better communication between different sub-communities within AI, improved publication practices, and more emphasis on reproducible research. They suggested that viewing AI research as a continuous spectrum rather than separate disciplines could lead to more impactful contributions. By fostering symbiotic interaction between theory-driven and application-driven approaches, the field can develop AI systems that are both powerful and scientifically grounded.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn