Researchers have developed a cost-effective and energy-saving robotic hand that can grasp a variety of objects, without dropping them, by utilizing only wrist movement and the feeling in its “skin.”
The ability to grasp objects of varying sizes, shapes, and textures is a task that is simple for humans but presents a challenge for robots. To address this issue, researchers from the University of Cambridge have created a flexible, 3D-printed robotic hand that, despite its inability to move its fingers independently, can still carry out a range of complex movements.
The robot hand was trained to grasp different objects and was able to predict whether it would drop them by using the information provided by sensors placed on its ‘skin’.
This type of passive movement makes the robot far easier to control and far more energy-efficient than robots with fully motorized fingers. The researchers say their adaptable design could be used in the development of low-cost robotics that are capable of more natural movement and can learn to grasp a wide range of objects. The results are reported in the journal Advanced Intelligent Systems.
In the natural world, movement results from the interplay between the brain and the body: this enables people and animals to move in complex ways without expending unnecessary amounts of energy. Over the past several years, soft components have begun to be integrated into robotics design thanks to advances in 3D printing techniques, which have allowed researchers to add complexity to simple, energy-efficient systems.
Robot hand picking up a peach using just the movement of its wrist. Credit: University of Cambridge
The human hand is highly complex, and recreating all of its dexterity and adaptability in a robot is a massive research challenge. Most of today’s advanced robots are not capable of manipulation tasks that small children can perform with ease. For example, humans instinctively know how much force to use when picking up an egg, but for a robot, this is a challenge: too much force and the egg could shatter; too little, and the robot could drop it. In addition, a fully actuated robot hand, with motors for each joint in each finger, requires a significant amount of energy.
In Professor Fumiya Iida’s Bio-Inspired Robotics Laboratory in Cambridge’s Department of Engineering, researchers have been developing potential solutions to both problems: a robot hand than can grasp a variety of objects with the correct amount of pressure while using a minimal amount of energy.
“In earlier experiments, our lab has shown that it’s possible to get a significant range of motion in a robot hand just by moving the wrist,” said co-author Dr. Thomas George-Thuruthel, who is now based at University College London (UCL) East. “We wanted to see whether a robot hand based on passive movement could not only grasp objects, but would be able to predict whether it was going to drop the objects or not, and adapt accordingly.”
The researchers used a 3D-printed anthropomorphic hand implanted with tactile sensors so that the hand could sense what it was touching. The hand was only capable of passive, wrist-based movement.
The team carried out more than 1200 tests with the robot hand, observing its ability to grasp small objects without dropping them. The robot was initially trained using small 3D-printed plastic balls and grasped them using a pre-defined action obtained through human demonstrations.
“This kind of hand has a bit of springiness to it: it can pick things up by itself without any actuation of the fingers,” said first author Dr Kieran Gilday, who is now based at EPFL in Lausanne, Switzerland. “The tactile sensors give the robot a sense of how well the grip is going, so it knows when it’s starting to slip. This helps it to predict when things will fail.”
The robot used trial and error to learn what kind of grip would be successful. After finishing the training with the balls, it then attempted to grasp different objects including a peach, a computer mouse, and a roll of bubble wrap. In these tests, the hand was able to successfully grasp 11 of 14 objects.
“The sensors, which are sort of like the robot’s skin, measure the pressure being applied to the object,” said George-Thuruthel. “We can’t say exactly what information the robot is getting, but it can theoretically estimate where the object has been grasped and with how much force.”
“The robot learns that a combination of a particular motion and a particular set of sensor data will lead to failure, which makes it a customizable solution,” said Gilday. “The hand is very simple, but it can pick up a lot of objects with the same strategy.”
“The big advantage of this design is the range of motion we can get without using any actuators,” said Iida. “We want to simplify the hand as much as possible. We can get lots of good information and a high degree of control without any actuators, so that when we do add them, we’ll get more complex behavior in a more efficient package.”
A fully actuated robotic hand, in addition to the amount of energy it requires, is also a complex control problem. The passive design of the Cambridge-designed hand, using a small number of sensors, is easier to control, provides a wide range of motion, and streamlines the learning process.
In the future, the system could be expanded in several ways, such as by adding computer vision capabilities or teaching the robot to exploit its environment, which would enable it to grasp a wider range of objects.
Reference: “Predictive Learning of Error Recovery with a Sensorized Passivity-Based Soft Anthropomorphic Hand” by Kieran Gilday, Thomas George-Thuruthel and Fumiya Iida, 12 April 2023, Advanced Intelligent Systems.
This work was funded by UK Research and Innovation (UKRI), and Arm Ltd. Fumiya Iida is a Fellow of Corpus Christi College, Cambridge.