Gemini Robotics-ER 1.6 lifts robot gauge reading to 93%
Original: We’re rolling out an upgrade designed to help robots reason about the physical world. 🤖 Gemini Robotics-ER 1.6 has significantly better visual and spatial understanding in order to plan and complete more useful tasks. Here’s why this is important 🧵 View original →
Google DeepMind’s new robotics thread matters because it ties a model update directly to physical-world tasks that industrial robots already struggle with: reading gauges, understanding multiple camera angles, and deciding whether a task is actually finished. In the source tweet, the company said Gemini Robotics-ER 1.6 is meant to help robots “reason about the physical world,” then used the rest of the thread to show where the upgrade lands.
“help robots reason about the physical world”
The clearest numbers came from the linked DeepMind blog post. On instrument reading, the company reports success rates of 23% for Gemini Robotics-ER 1.5, 67% for Gemini 3.0 Flash, 86% for Gemini Robotics-ER 1.6, and 93% when 1.6 uses agentic vision. The thread also claims the model is 10% better at detecting human injury risks in video, and better at respecting physical constraints such as not handling liquids or objects heavier than 20kg. Those details make the rollout more than a generic robotics refresh: DeepMind is positioning the model as a higher-level reasoning layer that can count, point, read instruments, fuse multiple live camera streams, and decide whether a plan step succeeded.
GoogleDeepMind’s account usually serves as a short bridge between polished launch material and developer entry points, and this thread follows that pattern closely. The post links to the blog, notes availability through the Gemini API and Google AI Studio, and points to a developer Colab. The blog adds useful context on why Boston Dynamics is in the picture: Spot already collects images from industrial facilities, and instrument reading is a real inspection task rather than a contrived benchmark. DeepMind also says the model can use code execution as part of “agentic vision,” which helps explain how it gets from a messy gauge image to a concrete reading.
What to watch next is whether developers outside Google’s partner set can reproduce these gains on their own hardware and camera setups. If the 93% instrument-reading result and the safety gains hold up beyond curated examples, Gemini Robotics-ER 1.6 could become a meaningful step toward robots that need less task-specific hand-tuning. Source tweet: GoogleDeepMind on X via Nitter.
Related Articles
Google DeepMind and Agile Robots said on March 24, 2026 that they will combine Gemini Robotics foundation models with Agile Robots hardware to build adaptable, reasoning robots for industrial environments. The partnership starts with high-value manufacturing and automation use cases where scale and reliability matter most.
r/singularity is less interested in the spectacle than in what 21 kilometres reveals about endurance, thermal limits, batteries, and autonomy. Euronews reports that more than 70 teams joined an overnight full-course test in Beijing E-Town ahead of the 2026-04-19 race, with around 40% now relying on fully autonomous navigation.
Google DeepMind has opened applications for a Europe-focused robotics accelerator. The three-month program combines technical mentorship, product support, and potential cloud credits up to $350k for eligible startups.
Comments (0)
No comments yet. Be the first to comment!