Fully Clickable Video Ad

Google’s Gemini Robotics AI Model Reaches Into the Physical World

Spread the love


In sci-fi tales, artificial intelligence often powers all sorts of clever, capable, and occasionally homicidal robots. A revealing limitation of today’s best AI is that, for now, it remains squarely trapped inside the chat window.

Google DeepMind signaled a plan to change that today—presumably minus the homicidal part—by announcing a new version of its AI model Gemini that fuses language, vision, and physical action together to power a range of more capable, adaptive, and potentially useful robots.

In a series of demonstration videos, the company showed several robots equipped with the new model, called Gemini Robotics, manipulating items in response to spoken commands: Robot arms fold paper, hand over vegetables, gently put a pair of glasses into a case, and complete other tasks. The robots rely on the new model to connect items that are visible with possible actions in order to do what they’re told. The model is trained in a way that allows behavior to be generalized across very different hardware.

Google DeepMind also announced a version of its model called Gemini Robotics-ER (for embodied reasoning), which has just visual and spatial understanding. The idea is for other robot researchers to use this model to train their own models for controlling robots’ actions.

Blinking Photo Ad

In a video demonstration, Google DeepMind’s researchers used the model to control a humanoid robot called Apollo, from the startup Apptronik. The robot converses with a human and moves letters around a tabletop when instructed to.

“We’ve been able to bring the world-understanding—the general-concept understanding—of Gemini 2.0 to robotics,” said Kanishka Rao, a robotics researcher at Google DeepMind who led the work, at a briefing ahead of today’s announcement.

See also  NASA Orbiter Captures Stunning View of Blue Ghost Lander on the Moon

Google DeepMind says the new model is able to control different robots successfully in hundreds of specific scenarios not previously included in their training. “Once the robot model has general-concept understanding, it becomes much more general and useful,” Rao said.

The breakthroughs that gave rise to powerful chatbots, including OpenAI’s ChatGPT and Google’s Gemini, have in recent years raised hope of a similar revolution in robotics, but big hurdles remain.

Related Posts
Kiren Rijiju: Why Earth Sciences minister Rijiju is upset with this European IT company | – Times of India

Earth Sciences Minister Kiren Rijiju is reportedly upset with the French IT company Atos. Reason is said to be Read more

Former Activision boss reportedly wants to buy TikTok – Times of India
Former Activision boss reportedly wants to buy TikTok - Times of India

Bobby Kotick, the former head of Activision Blizzard, is reportedly considering buying TikTok, as the app could be banned Read more

How Apple’s Find My app ‘cost’ a US city millions of dollars – Times of India
How Apple’s Find My app ‘cost’ a US city millions of dollars - Times of India

Apple's Find My app has cost the city of Denver, US $3.76 million in compensation and damages. In 2022, Read more

Moto G54 receives a price cut in India: Here’s how much the smartphone costs – Times of India
Moto G54 receives a price cut in India: Here’s how much the smartphone costs - Times of India

If you have been planing to purchase a budget smartphone, then you can consider buying the Moto G54. Launched Read more

Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top