AIResearchAIResearch
Machine Learning

DeepMind releases Gemini Robotics-ER 1.6 with spatial reasoning upgrades

Gemini Robotics-ER 1.6 adds relational reasoning, analog gauge reading, and a modular tool-calling layer for robots in factories, warehouses, and homes.

3 min read
DeepMind releases Gemini Robotics-ER 1.6 with spatial reasoning upgrades

TL;DR

Gemini Robotics-ER 1.6 adds relational reasoning, analog gauge reading, and a modular tool-calling layer for robots in factories, warehouses, and homes.

Spatial reasoning has always been robotics' hardest subproblem. Parsing "move the red block to the left of the tall bottle" requires simultaneous object recognition, scale comparison, and reference-frame inference -- all in real time. Google DeepMind's Gemini Robotics-ER 1.6, released Tuesday, targets this capability gap as a general-purpose reasoning layer for physical agents of all kinds.

According to SiliconAngle, the model introduces significant upgrades to multiview spatial understanding and precision object detection. Relational queries now work more reliably -- identifying the smallest item in a collection, handling directional movement instructions, or resolving constraints like "point to every object small enough to fit inside the blue cup." These may look like toy examples, but they represent exactly the kind of ambiguous natural-language input that real operators use in practice.

The model ships as a high-level orchestration layer rather than a standalone controller. It exposes native hooks into Google Search, interfaces with vision-language-action (VLA) models for low-level actuation, and accepts user-defined third-party functions. A warehouse integrator could connect a custom inventory database as a tool call without touching the underlying model weights.

Reading the gauges

Most robotics AI coverage focuses on manipulation and navigation. DeepMind's emphasis on instrument reading in this release is a notable signal about target deployment environments. Analog gauges -- needles against printed scales, overlapping tick marks, partially obscured displays -- present a complex visual reasoning challenge that earlier generalist vision models handled poorly. The company specifically cited this capability as enabling reliable operation in factories, warehouses, and domestic spaces where misread instrumentation carries real consequences.

Boston Dynamics' Spot is explicitly mentioned as a beneficiary. Inspection robots like Spot are routinely deployed to read instruments in environments too hazardous for regular human access, making gauge accuracy a safety-critical requirement rather than a benchmark metric.

Price Per Token and LLM Stats both logged the release alongside a crowded week of foundation model updates, with Gemini 3.1 Flash TTS and several other releases shipping in the same 48-hour window. The density of concurrent launches reflects how quickly the deployment pace has accelerated across major labs in 2026.

The architecture bet

DeepMind's design choice here -- a reasoning model that delegates action execution to specialized VLA models via tool calls -- encodes a specific thesis about how robotics AI will scale. Rather than training one model end-to-end from sensor input to motor command, the stack decomposes into a reasoning tier and an action tier, where the former handles language, context, and planning while the latter handles real-time control. This mirrors how software agents in language-only settings delegate to specialized APIs.

For practitioners, this pattern has clear appeal: updating the action model for a new robot platform does not require retraining the reasoning layer. The obvious risk is latency introduced by inter-component communication, and DeepMind has not yet published latency or reliability benchmarks for this integration path -- data engineers will need before committing to this architecture in production.

Google's investment in Gemini Robotics signals that the company treats embodied AI as a natural extension of its existing model platform rather than a parallel research track. The same multimodal foundation backing document understanding and code generation is now being tuned toward physical-world interaction. Whether that translates from benchmark demonstrations to reliable industrial deployment will determine if Gemini Robotics-ER 1.6 lands as infrastructure or remains a research proof-of-concept.

The real test will be messy factory floors and unpredictable domestic environments -- the conditions under which every assumption made in a clean lab demo tends to break.

---

Frequently asked questions

What is Gemini Robotics-ER 1.6?

Google DeepMind's updated foundation model for robotics, designed to serve as a high-level reasoning and planning layer. It handles task decomposition, tool calling, and spatial inference, then delegates motor execution to separate vision-language-action models.

How does the model handle relational instructions?

It can resolve comparative and constraint-based queries -- identifying the smallest object in a group, determining spatial from-to relationships, or finding items satisfying a size condition relative to another object -- without task-specific fine-tuning.

Which robots benefit from this release?

DeepMind positions Gemini Robotics-ER 1.6 as compatible with robots of any type. Boston Dynamics' Spot is explicitly cited, given its use in industrial inspection scenarios where reliable instrument reading matters most.

Why is reading analog gauges difficult for robotics AI?

Analog gauges combine needles, fine tick marks, printed scales, and contextual labels in a dense visual space. Correctly interpreting a reading requires understanding layout conventions and resolving fine-grained detail -- tasks that earlier generalist vision models handled inconsistently.

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn