AIResearch AIResearch
Back to articles
Data

AI Helps Robots See Boxes Better in Warehouses

A new method allows robots to accurately identify and handle boxes of any size without needing detailed 3D models, cutting processing time by 76% while maintaining precision.

AI Research
March 27, 2026
4 min read
AI Helps Robots See Boxes Better in Warehouses

In modern warehouses where robots increasingly handle logistics tasks, one fundamental has persisted: how can these machines accurately perceive and manipulate boxes they've never seen before? Traditional approaches either require precise 3D models for every box type—impractical for constantly changing inventories—or struggle with the geometric simplicity and visual similarity of cardboard containers. Researchers from Huawei Technologies Canada have developed a solution that bridges this gap, enabling robots to estimate both position and orientation of novel boxes using just a single camera and depth sensor observation.

The key finding is that Box6D, their proposed , achieves competitive or superior accuracy to existing approaches while dramatically reducing computation time. On proprietary warehouse data, the system achieved perfect precision at moderate overlap thresholds (1.00 at IoU=0.5 and 0.7) and maintained 0.92 precision at the strictest threshold (IoU=0.9), closely approaching the 0.95 performance of systems using ground-truth 3D models. Perhaps more impressively, Box6D reduced inference time by approximately 76% compared to standard approaches, making it suitable for real-time applications where speed matters as much as accuracy.

Ology cleverly combines several components to work with a single generic box template rather than requiring specific models for each box instance. First, an object detection module identifies boxes in the scene using segmentation techniques. Then, a pose estimation component generates multiple hypotheses about each box's position and orientation by matching point clouds between the observed scene and the template box. A critical innovation is the depth-consistency filter, which eliminates hypotheses where the rendered box geometry doesn't align with observed depth data—particularly important for symmetric boxes where incorrect rotations can appear plausible. Finally, a dimension estimation module uses binary search to determine the actual size of each box by comparing projected template masks with observed masks.

Across multiple datasets demonstrate the system's effectiveness. On the HouseCat6D benchmark, Box6D achieved 88.8 average precision at IoU=0.25 and 58.9 at IoU=0.50, substantially outperforming the next best (SecondPose at 54.5 and 23.7 respectively). On the PACE dataset focused on manipulation scenarios, Box6D achieved the best IoU scores (63.2 at IoU=0.25 and 10.4 at IoU=0.50) and highest average precision for rotation accuracy criteria. The researchers conducted ablation studies showing that the depth-consistency filter alone improved precision from 0.53 to 0.86 at IoU=0.80 on warehouse data, while the early-stopping mechanism reduced average runtime from 4.93 seconds to 1.16 seconds per frame while only slightly decreasing precision from 0.94 to 0.92.

For warehouse automation are significant. Current robotic systems often struggle with boxes due to their symmetrical shapes, weak textures, and frequent occlusion in cluttered environments. By enabling accurate pose estimation without requiring detailed 3D models for every box variant, Box6D could make robotic picking and placing more flexible and cost-effective. 's efficiency gains—achieving near real-time performance while maintaining accuracy—addresses a practical barrier to deployment in time-sensitive logistics operations where every second counts.

Despite these advances, limitations remain. is specifically tailored for box-shaped objects in warehouse contexts, so its applicability to other object categories or environments would require adaptation. The paper notes that while the approach handles symmetry and geometric ambiguity well through the depth-consistency filter, performance might degrade with extremely reflective surfaces or under severe lighting variations not represented in the training data. Additionally, the evaluation focuses on static or moderately dynamic scenes, leaving open questions about performance in highly dynamic environments with rapidly moving objects.

The researchers evaluated their system on three datasets spanning different s: proprietary warehouse data with diverse lighting and occlusion, HouseCat6D with household scenes and varied surface properties, and PACE with manipulation-centric scenarios. This comprehensive testing across environments with clutter, multiple instances, and different material properties provides confidence in 's robustness, though real-world deployment would likely reveal additional edge cases requiring attention. The paper's focus on practical industrial applications distinguishes it from more theoretical computer vision research, offering a pathway toward tangible improvements in automated logistics systems.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn