AIResearch AIResearch
Back to articles
Robotics

Robots Learn to Spot Vineyards with Minimal Human Help

A new AI system enables agricultural robots to detect and map vine trunks in real-time using limited training data, achieving over 70% accuracy in challenging field conditions.

AI Research
April 01, 2026
4 min read
Robots Learn to Spot Vineyards with Minimal Human Help

Agricultural robots are poised to transform farming by automating tasks like monitoring crops and predicting yields, but they often struggle to reliably identify plants in the unpredictable conditions of real-world fields. A new study introduces an AI framework that allows mobile robots to detect vine trunks with high accuracy, even when starting with very little labeled data. This advancement addresses a critical bottleneck in precision agriculture, where manual data annotation is labor-intensive and limits the deployment of autonomous systems in diverse environments.

The researchers developed a system that successfully identifies over 70% of vine trees in a single pass through vineyards, with a mean distance error of less than 0.37 meters. This performance was achieved in novel settings with varying lighting and crop densities, demonstrating robustness against common s like wind and vegetation interference. By integrating multi-modal sensors—including visible, near-infrared, and thermal cameras along with LiDAR—the framework builds a comprehensive understanding of the environment, enabling precise localization of tree trunks without relying on large, manually curated datasets.

Ology combines an annotation pipeline and a detection pipeline to train a robust multi-modal detector. Initially, the system uses a frozen semantic annotator, specifically the Segment Anything Model (SAM), to generate partial labels of vine trunks from images across different modalities. These labels are filtered based on shape criteria, such as rectangular contours, and then spatially associated using LiDAR point clouds to ensure consistency across sensors. This process creates pseudo-labels that enrich the training dataset with minimal human intervention, requiring only early-stage manual correction to remove false annotations.

In the detection pipeline, a pre-trained YOLOv10n model is employed in a multi-stage training procedure. Starting with a small dataset of 100 partially annotated image sets, the detector iteratively refines its predictions by incorporating high-confidence pseudo-labels from previous stages. As shown in Table I of the paper, the final detector, DT, achieved a precision of 0.83 and a recall of 0.53, a significant improvement from initial scores of 0.02 precision and 0.14 recall. Figure 5 illustrates this evolution, with the DT model producing accurate detections across all input modalities, prioritizing low false positives to ensure reliable automated labeling.

Experimental evaluations in real vineyard settings validated the system's effectiveness. In a single-row configuration with 10 trees, the detector identified 80% of the trees with a mean Euclidean distance error of 0.28 meters, as detailed in Table II. For a more complex dual-row setup with 14 trees, it detected 10 trees with a mean error of 0.37 meters, maintaining a recall rate above 70% despite variable sunlight conditions. Integration with a customized LiDAR and Odometry Mapping (LOAM) algorithm and a tree association module enabled the generation of feature-rich sparse maps, visualized in Figure 6, which combine thermal data with real-time trunk detections for comprehensive field assessment.

Of this work extend beyond vineyards to other agricultural and robotic applications where labeled data is scarce. By reducing the need for extensive human annotation, the framework makes it feasible to deploy autonomous systems in diverse, unstructured environments, potentially lowering labor costs and enhancing sustainability in farming. It also opens doors for scalable multi-agent detection systems that could fuse cross-modal information from multiple robots to further improve accuracy and efficiency in large-scale operations.

Despite its successes, the study acknowledges limitations, such as the trade-off between precision and recall, as the detector prioritizes minimizing false positives, which in lower mAP scores. Future work could explore integrating cross-agent data to maximize detection accuracy and further reduce human supervision. The researchers note that while the system performs well in tested conditions, its generalizability to other crop types or extreme environmental variations remains to be fully assessed, highlighting areas for continued investigation in real-world agricultural robotics.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn