Google DeepMind Unveils Gemini Robotics — New AI Models for Robotics

Google DeepMind has announced two new models based on Gemini 2.0, Gemini Robotics and Gemini Robotics-ER, which expand the robots' capabilities to perform real-world tasks.

Google DeepMind Unveils Gemini Robotics — New AI Models for Robotics

Google DeepMind has unveiled two new Gemini 2.0-based AI models for robotics. The models, Gemini Robotics and Gemini Robotics-ER, are designed to enhance robots' ability to perform real-world tasks.

Gemini Robotics is an advanced vision-language-action (VLA) model that adds physical actions to Gemini 2.0’s multimodal capabilities, allowing direct control of robots. Gemini Robotics-ER, on the other hand, offers improved spatial understanding, allowing roboticists to use Gemini’s embodied reasoning (ER) capabilities in their projects.

The company is collaborating with Apptronik to create a new generation of humanoid robots, and is also working with trusted testers to further develop the technology. These models open up new perspectives for applying AI in the physical world, making robots more useful and efficient.

To be useful, AI robots must be versatile (adaptable to different situations), interactive (quickly respond to changes), and dexterous (able to manipulate objects). Gemini Robotics takes a significant step forward in these areas, bringing the creation of versatile robots closer. Google emphasizes that the new models make it possible to create robots that can independently analyze and perform tasks in the physical world. This includes moving, solving logical problems, performing household functions, interacting with people, and analyzing changes in the environment.

What's Your Reaction?

like

dislike

love

funny

angry

sad

wow