Google researchers claim they have built a Transporter Network AI model architecture that allows object-grasping robots to reason about which visual signals are relevant and how they should be rearranged in a scene. The researchers say their Transporter Networks have achieved “superior” efficiency during experiments on a variety of tasks, including stacking a block pyramid, assembling kits, manipulating ropes, and moving small object piles. 

Grasping robotics is a challenge. Robots, for instance, fail to do what is called “mechanical search,” which is when they have to locate an object from within a stack of other objects and pick it up. Most robots are not particularly adaptable, and there is a shortage of sufficiently capable AI models in mechanical quest to direct robot pincers, an issue that has come to the fore as the pandemic causes businesses to consider implementing automation.

The Google study coauthors say Transporter Networks don’t need any prior 3D model, pose, or class category knowledge of the objects to be manipulated, instead relying only on information found within partial depth camera data. They’re also capable of generalizing to new artifacts and configurations and, for certain activities, learning from a single demonstration. In reality, Transporter Networks, trained from scratch, ostensibly achieved over 90% performance on most tasks with artifacts in new configurations on 10 specific tabletop manipulation tasks using 100 expert video demonstrations of the tasks.

On datasets of demonstrations ranging in number from one demonstration to 1,000 per mission, the researchers trained Transporter Networks. They first deployed them on Ravens, a virtual benchmark learning environment with a suction gripper overlooking a 0.5 x 1 meter workspace consisting of a Universal Robot UR5e device. Then, using true UR5e robots with suction grippers and cameras including an Azure Kinect, they validated the Transporter Networks on package assembly tasks.

The researchers conducted their experiments by using a Unity-based software that allows individuals to remotely teleoperate robots due to pandemic-related lockdowns. The teleoperators were charged with assembling and disassembling a package of five small bottled mouthwashes or nine uniquely formed wooden toys repeatedly for one experiment, using either a virtual reality headset or mouse and keyboard to mark picking and placing poses. The Transporter Networks, which were trained on all tasks by 13 human operators with 11,633 pick-and – place acts in total, achieved 98.9 percent success in assembling bottled mouthwash kits.

“We presented the Transporter Network in this work, a simple model architecture that infers spatial displacements that can parameterize visual input robot behavior,” the researchers wrote. It allows no object assumptions, exploits spatial symmetries, and is more effective in learning vision-based manipulation tasks in order of magnitude than end-to – end alternatives … In terms of its current limitations: it is responsive to camera-robot tuning, and how to combine torque and force behavior with spatial action spaces remains uncertain. Overall, we are enthusiastic about this course and plan to expand it to high-rate monitoring in real-time, as well as tool-use activities. 

The coauthors state they plan to release Ravens (and an associated API) code and open-source in the near future.

Comments to: Google’s Transporter Network uses minimal data to teach robots to stack blocks

Your email address will not be published. Required fields are marked *

Attach images - Only PNG, JPG, JPEG and GIF are supported.

Good Reads

When urban development takes place, a traffic impact assessment is often needed before a project is approved: What will happen to auto traffic if a new apartment building or business complex is constructed, or if a road is widened? On the other hand, new developments affect foot traffic as well — and yet few places study the effects of urban change on pedestrians.

Worlwide

When urban development takes place, a traffic impact assessment is often needed before a project is approved: What will happen to auto traffic if a new apartment building or business complex is constructed, or if a road is widened? On the other hand, new developments affect foot traffic as well — and yet few places study the effects of urban change on pedestrians.
This week, The European Parliament, the body responsible for adopting European Union (EU) legislation, passed a non-binding resolution calling for a ban on law enforcement use of facial recognition technology in public places. The resolution, which also proposes a moratorium on the deployment of predictive policing software, would restrict the use of remote biometric identification unless it’s to fight “serious” crime, such as kidnapping and terrorism.

Trending

WHEN SARTRE SAID hell is other people, he wasn’t living through 2020. Right now, other people are the only thing between us and species collapse. Not just the people we occasionally encounter behind fugly masks—but the experts and innovators out in the world, leading the way. The 17-year-old hacker building his own coronavirus tracker. The […]
13 September marks six months since the first coronavirus announced in Ethiopia.In the half-year since then, reported cases are close to 64 Thousend, with more than 996 deaths. At the onset, COVID-19 mainly affected the capital city. However, the virus is now moving from high-density urban areas to informal settlements and then onward to rural […]
Present international artificial intelligence (AI) inventory and progression in self-driving vehicle research and development Complementary subjects in technology are also artificial intelligence ( AI) and self-driving vehicles. In brief, without someone involved, you just can’t debate one. While AI has been rapidly applied in different areas, a new hot topic has been the way you […]

Login

Welcome to Intech Analytica

AI news hub. It checks trusted sites and collects best pieces of AI info.
Join Intech Analytica

Agree the terms of use.