AIResearch AIResearch
Back to articles
Science

AI's Culinary Breakthrough: How LLMs Are Revolutionizing Food Recognition

In the bustling world of artificial intelligence, a new study is serving up a fresh approach to a long-standing problem: teaching machines to recognize food accurately in real-world settings. Research…

AI Research
November 22, 2025
4 min read
AI's Culinary Breakthrough: How LLMs Are Revolutionizing Food Recognition

In the bustling world of artificial intelligence, a new study is serving up a fresh approach to a long-standing problem: teaching machines to recognize food accurately in real-world settings. Researchers from Singapore Management University have developed a framework that leverages large language models (LLMs) to tackle the trifecta of s in food recognition—domain shifts, long-tailed data distributions, and fine-grained categorization. Published in a recent arXiv preprint, this uses AI to generate textual descriptions of food images, bridging gaps between idealized internet-sourced photos and messy, user-captured snapshots from daily life. By aligning visual and textual features in a shared embedding space, the system not only improves accuracy but also offers a scalable solution for applications like dietary monitoring and smart food logging, potentially transforming how we interact with nutrition technology.

To understand ology, it's essential to grasp the core issues addressed. The researchers focused on datasets like FoodAI-HPB and UPMC-ETH Food101, where source domain images are crawled from the web and target domain images come from free-living environments via mobile apps. This setup creates a domain shift, where factors like lighting, background, and food presentation differ drastically between domains. Additionally, these datasets exhibit long-tailed distributions, meaning some food categories have abundant samples while others are severely underrepresented, and fine-grained distinctions make it hard to differentiate visually similar dishes, such as chicken rice versus boiled kampung chicken. The proposed framework begins by using BLIP2, an LLM, to parse food images and generate titles (e.g., "milk") and ingredients (e.g., "rice, cucumber, chicken") through specific prompts. These texts are encoded with hierarchical Transformers, while images are processed via ResNet-50, and both are projected into a shared space using a bi-directional triplet loss for alignment. This cross-modal approach ensures that images from the same category across domains are pulled closer together, while a calibration-based loss handles class imbalance by adjusting for sample frequencies, resulting in a robust model trained with minimal computational overhead.

From extensive experiments demonstrate the efficacy of this LLM-powered approach. On the FoodAI-HPB dataset, achieved a Top-1 accuracy of 60.7% and a Top-5 accuracy of 87.2%, outperforming numerous state-of-the-art techniques in domain adaptation, imbalance learning, and fine-grained recognition. For instance, it surpassed s like LTDS and BoDA, which combine domain alignment with imbalance handling but often struggle with fine-grained categories. Similarly, on the UPMC-ETH Food101 dataset, it reached a Top-1 accuracy of 85.6% and a Top-5 accuracy of 96.9%, with particularly impressive gains in tail classes—showing a 53% improvement over baseline models. Ablation studies confirmed that each component, from alignment loss to text augmentation with titles and ingredients, contributed incrementally to performance. Visualizations using t-SNE plots revealed tighter feature clusters and clearer class boundaries compared to ERM baselines, highlighting how LLM-generated texts help separate fine-grained categories and support underrepresented classes without sacrificing overall accuracy.

Of this research extend far beyond academic benchmarks, promising real-world impacts in healthcare, consumer technology, and beyond. By enabling more accurate food recognition in diverse environments, this approach could enhance mobile apps for dietary assessment, helping users track nutrition in real-time with greater reliability. It addresses practical scenarios where data is imbalanced and domains vary, making it suitable for global applications that account for cultural and geographical food variations. Moreover, the use of LLMs for cross-modal alignment introduces a low-cost augmentation strategy that could be adapted to other vision tasks, such as medical imaging or retail, where textual context can clarify ambiguous visuals. However, the study also underscores the importance of LLM reliability; incorrect text generations, especially for geography-specific foods, can hinder performance, suggesting a need for validation mechanisms in future deployments to ensure robustness across different populations and cuisines.

Despite its successes, the approach has limitations that warrant consideration. The dependency on LLMs like BLIP2 means that performance can degrade if the model fails to generate accurate texts for niche or regional dishes, as seen in the FoodAI-HPB dataset where BLIP2's Top-1 accuracy dropped to 13.9%. Comparisons with other LLMs such as LLaVA and FoodLMM showed that overly detailed texts incorporating tastes or origins could complicate alignment, indicating that generality in text generation is crucial. Additionally, the framework assumes shared categories across domains, which may not hold in all real-world cases, and it does not fully address scenarios where source domain samples are scarce. Future work could explore dynamic text validation, multi-modal fusion techniques, or applications in other long-tailed domain adaptation problems to build on these foundations and overcome these constraints.

Reference: Wang, Q., Ngo, C.-W., Lim, E.-P., & Sun, Q. (2025). LLMs-based Augmentation for Domain Adaptation in Food Datasets. arXiv preprint arXiv:2511.16037v1.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn