Site iconSite icon ForkLog

Former Google Engineers Unveil AI for Robots with ‘Untrained’ Skills

Former Google Engineers Unveil AI for Robots with 'Untrained' Skills

A startup founded by former Google engineers, Physical Intelligence, has unveiled the π0.7 model. The developers have claimed a “qualitative leap” in the AI’s ability to generalize skills and perform tasks it was not directly trained for.

The system belongs to the “Vision-Language-Action” (VLA) class and is designed for robot control.

Unlike previous solutions, π0.7 has demonstrated signs of compositional generalization—the ability to combine previously learned skills to solve new tasks.

Untrained Tasks and Transfer Between Robots

During experiments, the model exhibited a range of unexpected abilities. Notably, π0.7 was able to control a new type of robot and fold t-shirts, despite the absence of training data for this specific platform.

The results are comparable to the level of operators with hundreds of hours of teleoperation experience, noted the programmers.

The tool also managed to understand the use of previously unfamiliar devices, including kitchen appliances. For instance, the robot completed part of a task involving cooking sweet potatoes in an air fryer, even though such scenarios were not in the training set.

According to the developers, this was made possible by combining disparate skills—similar to how language models combine knowledge from different domains.

Control Through Language and Context

One of the key differences of π0.7 is its ability to be controlled not only through “what to do” commands but also through “how to do” clarifications.

The model accepts:

Some of the subgoals can be created by the auxiliary system during operation. This allows the robot to adjust its behavior without retraining.

This approach allows for the integration of data from various sources—video, telemetry from robots, and autonomously collected episodes—into a unified learning system.

The First Step Towards ‘Universal’ Robots

Physical Intelligence noted that previously such models required retraining for each task—similar to early versions of language models. In contrast, π0.7 works “out of the box” and adapts to new scenarios through language.

The team emphasized that such a level of generalization has long been considered a strong point of LLM, but remained unattainable in robotics.

Despite the progress, the model still struggles with complex tasks without step-by-step prompts. However, with sequential instructions, the quality of execution significantly improves.

In the future, such instructions could help train more autonomous machines capable of acting without human intervention. Physical Intelligence believes that π0.7 shows the first signs of a transition to universal robots that adapt to new conditions without manual adjustment for each task.

Back in February, the company Carbon Robotics released the AI model Large Plant Model, which can recognize plant species to combat weeds.

Exit mobile version