AIResearch AIResearch
Back to articles
AI

AI Spots Appendicitis in Kids' Ultrasounds

Deep learning model achieves 93% accuracy detecting pediatric appendicitis from ultrasound images, potentially reducing diagnostic errors and unnecessary surgeries

AI Research
November 11, 2025
3 min read
AI Spots Appendicitis in Kids' Ultrasounds

Appendicitis remains the most common surgical emergency in children, yet accurate diagnosis continues to challenge clinicians due to overlapping symptoms and variable ultrasound quality. A new artificial intelligence approach could help address this persistent medical problem by providing consistent, objective analysis of ultrasound images.

Researchers have developed a deep learning system that can accurately identify appendicitis in pediatric ultrasound images with 93.44% overall accuracy. The model specifically achieved 89.8% sensitivity (recall) for detecting true cases of appendicitis and 91.53% precision, meaning it correctly identified most cases while minimizing false alarms. This performance represents a significant improvement over traditional diagnostic methods that rely on subjective human interpretation.

The team used a convolutional neural network architecture called ResNet-50, which was pre-trained on general image recognition tasks and then fine-tuned specifically for appendicitis detection. They trained their model on the Regensburg Pediatric Appendicitis Dataset, which contains ultrasound scans from St. Hedwig Hospital in Germany collected between 2016 and 2021. Each patient record included up to 15 different ultrasound views covering the right lower quadrant, lymph nodes, and related anatomical structures.

Before analysis, the ultrasound images underwent preprocessing including normalization, resizing to 224x224 pixels, and data augmentation techniques like rotation and contrast adjustments to improve generalization. The model was trained using 80% of the data and tested on the remaining 20%, with performance evaluated using standard metrics including accuracy, precision, recall, F1-score, and area under the ROC curve (AUC).

The results showed the model achieved an AUC of 0.95, indicating excellent ability to distinguish between appendicitis and non-appendicitis cases. Analysis of the model's decision-making process using Gradient-weighted Class Activation Mapping (Grad-CAM) revealed that the AI system focused on anatomically relevant areas that radiologists typically examine, including the appendiceal region, pericecal fat, and bowel loops. This alignment with clinical practice suggests the model learned medically meaningful patterns rather than superficial artifacts.

From a clinical perspective, the high sensitivity is particularly valuable for reducing missed diagnoses of appendicitis, which can lead to serious complications if left untreated. The strong precision also helps minimize false positive diagnoses that might otherwise result in unnecessary appendectomies. In resource-constrained emergency departments, such AI assistance could provide standardized support for clinicians of varying experience levels and help reduce diagnostic variability.

The study does have limitations. The dataset, while comprehensive, is relatively modest by deep learning standards, and the model was only tested on static B-mode ultrasound images rather than video sequences. Differences in ultrasound machines and operator experience across institutions could also affect performance when applied in new clinical settings. Future work should explore combining imaging analysis with clinical data like white blood cell counts and C-reactive protein levels for even more accurate diagnosis.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn