AIResearch AIResearch
Back to articles
Robotics

Robots Learn to Uncover Hidden Objects Efficiently

A new AI method helps robots find and retrieve items buried in clutter by combining teacher guidance and visual cues, improving graspability without complex models.

AI Research
November 14, 2025
3 min read
Robots Learn to Uncover Hidden Objects Efficiently

In cluttered environments like warehouses or homes, robots often struggle to retrieve objects that are partially hidden under piles of items. This challenge, known as Mechanical Search, requires robots to interact with their surroundings to uncover targets, but traditional methods rely on slow, pre-programmed actions that can fail in unpredictable settings. A new study introduces a deep reinforcement learning approach that enables robots to learn efficient pushing strategies, significantly improving their ability to expose and grasp objects without extensive trial and error.

The researchers developed a method where a robot uses visual inputs from an RGB-D camera to guide its actions in real time. Instead of relying solely on raw images, the system processes a mid-level representation that highlights the approximate position of the target object through segmentation masks. This input, combined with the robot's end-effector position, allows the policy to generate precise pushing motions. The key finding is that this approach, enhanced by teacher-guided exploration and privileged information during training, leads to faster learning and better performance in uncovering objects compared to baseline methods.

To train the robot, the team employed a combination of algorithmic strategies in simulation. First, they used teacher policies—expert-designed actions like straight-line pushes or spiral motions—to guide the reinforcement learning agent during early stages, reducing the number of interactions needed. Second, they incorporated privileged information, such as the exact positions of objects, which was only available in simulation to speed up training. The agent itself was trained using a variant of the Deep Deterministic Policy Gradient (DDPG) algorithm, focusing on continuous control actions that adjust the end-effector's position based on current observations. The reward function encouraged uncovering the target while penalizing unnecessary movements of the heap or the target itself, ensuring efficient and safe interactions.

Experiments in simulation with single-heap and dual-heap setups showed that the method converged to effective solutions within about 8,000 interaction steps. In the single-heap condition, the agent achieved a 23% average improvement in graspability—measured by a grasp quality network—and successfully uncovered the target object in most cases. For the more challenging dual-heap scenario, where objects are spread across two piles, the approach still performed well, though it highlighted the importance of the mid-level representation for handling uncertainty. Ablation studies confirmed that removing components like teacher guidance or the asymmetric actor-critic setup led to slower learning and poorer results, emphasizing the necessity of the combined strategies.

This advancement matters because it addresses a common problem in robotics applications, from logistics to domestic assistance, where quick and reliable object retrieval is crucial. By enabling robots to adapt to dynamic environments without precise models, the method reduces execution times and minimizes unwanted disturbances, making automated systems more practical and efficient. However, limitations remain: the policy sometimes fails in cases where it repeats actions without effect or avoids interaction altogether, as noted in the paper's failure modes. Future work could focus on refining the agent's architecture to overcome these issues and testing the approach in real-world settings to ensure robustness beyond simulation.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn