Gemini Robotics

Vision-language-action model

Gemini Robotics is an advanced vision-language-action model developed by Google DeepMind[1] in partnership with Apptronik.[2] It is based on the Gemini 2.0 large language model.[3] It is tailored for robotics applications and can understand new situations.[4][5] There is a related version called Gemini Robotics-ER, which stands for embodied reasoning.[3] The two models were launched on March 12, 2025.[5]

On June 24, 2025, Google DeepMind released Gemini Robotics On-Device, a variant designed and optimized to run locally on robotic devices.[6]

Access to Gemini Robotics models is currently restricted to trusted testers, including Agile Robots, Agility Robotics,[7] Boston Dynamics, and Enchanted Tools.[2]

References

  1. ^ "Gemini Robotics". deepmind.google. Retrieved March 12, 2025.
  2. ^ a b Parada, Carolina. "Gemini Robotics brings AI into the physical world". Google DeepMind. Retrieved July 11, 2025.
  3. ^ a b Knight, Will (May 12, 2025). "Google's Gemini Robotics AI Model Reaches Into the Physical World". WIRED. Retrieved March 12, 2025.
  4. ^ "Google introduces new AI models for rapidly growing robotics industry". Reuters. March 12, 2025. Retrieved March 12, 2025.
  5. ^ a b Roth, Emma (March 12, 2025). "Google DeepMind's new AI models help robots perform physical tasks, even without training". The Verge. Retrieved March 12, 2025.
  6. ^ Parada, Carolina. "Gemini Robotics On-Device brings AI to local robotic devices". Google DeepMind. Retrieved July 11, 2025.
  7. ^ "Training a Whole-Body Control Foundation Model". Agility Robotics. Archived from the original on September 28, 2025. Retrieved January 6, 2026.


Retrieved from "https://en.wikipedia.org/w/index.php?title=Gemini_Robotics&oldid=1331465367"