Researchers develop new system with robot hands that learns how to grasp objects

Cluster of Excellence CITEC presents new system that learns how to grasp objects

Researchers at Bielefeld University have developed a grasp system with robot hands that autonomously familiarizes itself with novel objects. The new system works without previously knowing the characteristics of objects, such as pieces of fruit or tools. It was developed as part of the large-scale research project Famula at Bielefeld University's Cluster of Excellence Cognitive Interaction Technology (CITEC). The knowledge gained from this project could contribute to future service robots, for instance, that are able to independently adapt to working in new households. CITEC has invested approximately one million Euro in Famula. In a new "research_tv" report from Bielefeld University, the coordinators of the Famula project explain the new innovation.

"Our system learns by trying out and exploring on its own - just as babies approach new objects," says neuroinformatics Professor Dr. Helge Ritter, who heads the Famula project together with sports scientist and cognitive psychologist Professor Dr. Thomas Schack and robotics Privatdozent Dr. Sven Wachsmuth.

The CITEC researchers are working on a robot with two hands that are based on human hands in terms of both shape and mobility. The robot brain for these hands has to learn how everyday objects like pieces of fruit, dishes, or stuffed animals can be distinguished on the basis of their color or shape, as well as what matters when attempting to grasp the object.

The Human Being as the Model
A banana can be held, and a button can be pressed. "The system learns to recognize such possibilities as characteristics, and constructs a model for interacting and re-identifying the object," explains Ritter.

To accomplish this, the interdisciplinary project brings together work in artificial intelligence with research from other disciplines. Thomas Schack's research group, for instance, investigated which characteristics study participants perceived to be significant in grasping actions. In one study, test subjects had to compare the similarity of more than 100 objects. "It was surprising that weight hardly plays a role. We humans rely mostly on shape and size when we differentiate objects," says Thomas Schack. In another study, test subjects' eyes were covered and they had to handle cubes that differed in weight, shape, and size. Infrared cameras recorded their hand movements. "Through this, we find out how people touch an object, and which strategies they prefer to use to identify its characteristics," explains Dirk Koester, who is a member of Schack's research group. "Of course, we also find out which mistakes people make when blindly handling objects."

System Puts Itself in the Position of Its "Mentor"
Dr. Robert Haschke, a colleague of Helge Ritter, stands in front of a large metal cage with both robot arms and a table with various test objects. In his role as a human learning mentor, Dr. Haschke helps the system to acquire familiarity with novel objects, telling the robot hands which object on the table they should inspect next. To do this, Haschke points to individual objects, or gives spoken hints, such as in which direction an interesting object for the robot can be found (e.g. "behind, at left"). Using color cameras and depth sensors, two monitors display how the system perceives its surroundings and reacts to instructions from humans.

"In order to understand which objects they should work with, the robot hands have to be able to interpret not only spoken language, but also gestures," explains Sven Wachsmuth, of CITEC's Central Labs. "And they also have to be able to put themselves in the position of a human to also ask themselves if they have correctly understood." Wachsmuth and his team are not only responsible for the system's language capabilities: they have also given the system a face. From one of the monitors, Flobi follows the movements of the hands and reacts to the researchers' instructions. Flobi is a stylized robot head that complements the robot's language and actions with facial expressions. As part of the Famula system, the virtual version of the robot Flobi is currently in use.

Understanding Human Interaction
With the Famula project, CITEC researchers are conducting basic research that can benefit self-learning robots of the future in both the home and industry. "We want to literally understand how we learn to 'grasp' our environment with our hands. The robot makes it possible for us to test our findings in reality and to rigorously expose the gaps in our understanding. In doing so, we are contributing to the future use of complex, multi-fingered robot hands, which today are still too costly or complex to be used, for instance, in industry," explains Ritter.​

Comments

The opinions expressed here are the views of the writer and do not necessarily reflect the views and opinions of News Medical.
Post a new comment
Post

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.

You might also like...
Inside the Alzheimer's Association: Dr. Heather Snyder on Driving Research and Collaboration