Subscribe to our free daily newsletters
  Robot Technology News  




Subscribe to our free daily newsletters



ROBO SPACE
Robots foresee future with automated visualized predictions
by Staff Writers
Berkeley CA (SPX) Dec 12, 2017


UC Berkeley researchers have developed a robotic learning technology that enables robots to imagine the future of their actions so they can figure out how to manipulate objects they have never encountered before. In the future, this technology could help self-driving cars anticipate future events on the road and produce more intelligent robotic assistants in homes, but the initial prototype focuses on learning simple manual skills entirely from autonomous play. Watch a video explanation here

University of California, Berkeley, researchers have developed a robotic learning technology that enables robots to imagine the future of their actions so they can figure out how to manipulate objects they have never encountered before. In the future, this technology could help self-driving cars anticipate future events on the road and produce more intelligent robotic assistants in homes, but the initial prototype focuses on learning simple manual skills entirely from autonomous play.

Using this technology, called visual foresight, the robots can predict what their cameras will see if they perform a particular sequence of movements. These robotic imaginations are still relatively simple for now - predictions made only several seconds into the future - but they are enough for the robot to figure out how to move objects around on a table without disturbing obstacles.

Crucially, the robot can learn to perform these tasks without any help from humans or prior knowledge about physics, its environment or what the objects are. That's because the visual imagination is learned entirely from scratch from unattended and unsupervised exploration, where the robot plays with objects on a table. After this play phase, the robot builds a predictive model of the world, and can use this model to manipulate new objects that it has not seen before.

"In the same way that we can imagine how our actions will move the objects in our environment, this method can enable a robot to visualize how different behaviors will affect the world around it," said Sergey Levine, assistant professor in Berkeley's Department of Electrical Engineeing and Computer Sciences, whose lab developed the technology. "This can enable intelligent planning of highly flexible skills in complex real-world situations."

The research team will perform a demonstration of the visual foresight technology at the Neural Information Processing Systems conference in Long Beach, California, on December 5.

At the core of this system is a deep learning technology based on convolutional recurrent video prediction, or dynamic neural advection (DNA). DNA-based models predict how pixels in an image will move from one frame to the next based on the robot's actions. Recent improvements to this class of models, as well as greatly improved planning capabilities, have enabled robotic control based on video prediction to perform increasingly complex tasks, such as sliding toys around obstacles and repositioning multiple objects.

"In that past, robots have learned skills with a human supervisor helping and providing feedback. What makes this work exciting is that the robots can learn a range of visual object manipulation skills entirely on their own," said Chelsea Finn, a doctoral student in Levine's lab and inventor of the original DNA model.

With the new technology, a robot pushes objects on a table, then uses the learned prediction model to choose motions that will move an object to a desired location. Robot use the learned model from raw camera observations to teach themselves how to avoid obstacles and push objects around obstructions.

"Humans learn object manipulation skills without any teacher through millions of interactions with a variety of objects during their lifetime. We have shown that it possible to build a robotic system that also leverages large amounts of autonomously collected data to learn widely applicable manipulation skills, specifically object pushing skills," said Frederik Ebert, a graduate student in Levine's lab who worked on the project.

Since control through video prediction relies only on observations that can be collected autonomously by the robot, such as through camera images, the resulting method is general and broadly applicable.

In contrast to conventional computer vision methods, which require humans to manually label thousands or even millions of images, building video prediction models only requires unannotated video, which can be collected by the robot entirely autonomously. Indeed, video prediction models have also been applied to datasets that represent everything from human activities to driving, with compelling results.

"Children can learn about their world by playing with toys, moving them around, grasping, and so forth. Our aim with this research is to enable a robot to do the same: to learn about how the world works through autonomous interaction," Levine said. "The capabilities of this robot are still limited, but its skills are learned entirely automatically, and allow it to predict complex physical interactions with objects that it has never seen before by building on previously observed patterns of interaction."

The Berkeley scientists are continuing to research control through video prediction, focusing on further improving video prediction and prediction-based control, as well as developing more sophisticated methods by which robots can collected more focused video data, for complex tasks such as picking and placing objects and manipulating soft and deformable objects such as cloth or rope, and assembly.

Watch a video explanation here

ROBO SPACE
Toyota showcases humanoid robot that mirrors user
Tokyo (AFP) Nov 29, 2017
Japanese auto giant Toyota Wednesday showcased a humanoid robot that can mirror its user's movements, a product it says has uses as varied as elderly care and disaster response. The T-HR3 can be controlled by a wearable system that allows users to operate the entire robot in real-time by simply moving their own limbs. It is one of the main attractions at the International Robot Exhibitio ... read more

Related Links
University of California - Berkeley
All about the robots on Earth and beyond!


Thanks for being here;
We need your help. The SpaceDaily news network continues to grow but revenues have never been harder to maintain.

With the rise of Ad Blockers, and Facebook - our traditional revenue sources via quality network advertising continues to decline. And unlike so many other news sites, we don't have a paywall - with those annoying usernames and passwords.

Our news coverage takes time and effort to publish 365 days a year.

If you find our news sites informative and useful then please consider becoming a regular supporter or for now make a one off contribution.

SpaceDaily Contributor
$5 Billed Once


credit card or paypal
SpaceDaily Monthly Supporter
$5 Billed Monthly


paypal only

Comment using your Disqus, Facebook, Google or Twitter login.

Share this article via these popular social media networks
del.icio.usdel.icio.us DiggDigg RedditReddit GoogleGoogle

ROBO SPACE
Hensoldt intros new counter-drone system

China says Indian drone 'invaded' its airspace, crashed

Falcon's attack strategy could inspire new drones: study

'Go home' drone seeks to stop Japan overtime binge

ROBO SPACE
Nature's toughest substances decoded

Russia says 'satellite' could have caused radioactive pollution

NASA Selects Three Companies to Develop 'FabLab' Prototypes

US has lost dominance in highly intense, ultrafast laser technology to Europe and Asia

ROBO SPACE
Toshiba, Western Digital settle legal battle over chip unit sale

Secure information transmission over 500m fiber links based on quantum technologies

Squeezing light into a tiny channel brings optical computing a step closer

Researchers quantify factors for reducing power semiconductor resistance by two-thirds

ROBO SPACE
Mainz physicists propose a new method for monitoring nuclear waste

Australian waste treatment technology plays major role in management of radioactive waste

AREVA NP Maintenance Technique Reduces Frequency of Component Inspections at Two Exelon Plants

Cairo, Moscow sign contract for Egypt's first nuclear plant

ROBO SPACE
US air strike destroys vehicle bomb outside Mogadishu

Six Nigerian soldiers killed in Boko Haram ambushes

IS battle may be won, but Iraq faces major challenges

'Defeated' IS still poses danger in Iraq

ROBO SPACE
US void hard to miss at Paris climate summit

To save climate, stop investing in fossil fuels: economists

Improving sensor accuracy to prevent electrical grid overload

Japan faces challenges in cutting CO2, Moody's finds

ROBO SPACE
Nuclear fusion project faces delay over US budget cuts: director

Scientists create stretchable battery made entirely out of fabric

Surrey scientists create cheap and safe electro-catalysts for fuel cells

Army researchers seek better batteries

ROBO SPACE
Nation 'leads world' in remote sensing technology

China plans for nuclear-powered interplanetary capacity by 2040

China plans first sea based launch by 2018

China's reusable spacecraft to be launched in 2020




Memory Foam Mattress Review
Newsletters :: SpaceDaily :: SpaceWar :: TerraDaily :: Energy Daily
XML Feeds :: Space News :: Earth News :: War News :: Solar Energy News






The content herein, unless otherwise known to be public domain, are Copyright 1995-2017 - Space Media Network. All websites are published in Australia and are solely subject to Australian law and governed by Fair Use principals for news reporting and research purposes. AFP, UPI and IANS news wire stories are copyright Agence France-Presse, United Press International and Indo-Asia News Service. ESA news reports are copyright European Space Agency. All NASA sourced material is public domain. Additional copyrights may apply in whole or part to other bona fide parties. All articles labeled "by Staff Writers" include reports supplied to Space Media Network by industry news wires, PR agencies, corporate press officers and the like. Such articles are individually curated and edited by Space Media Network staff on the basis of the report's information value to our industry and professional readership. Advertising does not imply endorsement, agreement or approval of any opinions, statements or information provided by Space Media Network on any Web page published or hosted by Space Media Network. Privacy Statement