Particle.news

Download on the App Store

DeepMind Launches Gemini Robotics 1.5 for Web-Informed, Multi-Step Robot Tasks

The split design separates high-level reasoning from motor control to bring web-informed planning to physical tasks.

Overview

  • Google DeepMind announced Gemini Robotics-ER 1.5 for embodied reasoning alongside Gemini Robotics 1.5 as the action model.
  • The ER planner perceives the environment and can call web tools like Google Search to generate step-by-step natural-language instructions.
  • The action model executes those instructions with visual feedback, enabling multi-stage tasks such as sorting laundry, packing a suitcase using current London weather, and location-specific recycling.
  • DeepMind reported cross-embodiment transfer, with skills moving from the ALOHA2 platform to Franka arms and Apptronik’s Apollo humanoid without specialized tuning.
  • Developers can access ER 1.5 via the Gemini API in Google AI Studio, while the control model remains limited to trusted partners and testers.