Liquid Metal Sensors and AI Could Help Prosthetic Hands to ‘Feel’
Researchers used individual fingertips fitted with stretchable tactile sensors with liquid metal on a prosthesis attached to a robotic arm. (Photo by Alex Dolce)
Each fingertip has more than 3,000 touch receptors, which largely respond to pressure. Humans rely heavily on sensation in their fingertips when manipulating an object. The lack of this sensation presents a unique challenge for individuals with upper limb amputations. While there are several high-tech, dexterous prosthetics available today – they all lack the sensation of “touch.” The absence of this sensory feedback results in objects inadvertently being dropped or crushed by a prosthetic hand.
To enable a more natural feeling prosthetic hand interface, researchers from ’s and collaborators are the first to incorporate stretchable tactile sensors using liquid metal on the fingertips of a prosthetic hand. Encapsulated within silicone-based elastomers, this technology provides key advantages over traditional sensors, including high conductivity, compliance, flexibility and stretchability. This hierarchical multi-finger tactile sensation integration could provide a higher level of intelligence for artificial hands.
For the study, published in the journal , researchers used individual fingertips on the prosthesis to distinguish between different speeds of a sliding motion along different textured surfaces. The four different textures had one variable parameter: the distance between the ridges. To detect the textures and speeds, researchers trained four machine learning algorithms. For each of the ten surfaces, 20 trials were collected to test the ability of the machine learning algorithms to distinguish between the ten different complex surfaces comprised of randomly generated permutations of four different textures.
Results showed that the integration of tactile information from liquid metal sensors on four prosthetic hand fingertips simultaneously distinguished between complex, multi-textured surfaces – demonstrating a new form of hierarchical intelligence. The machine learning algorithms were able to distinguish between all the speeds with each finger with high accuracy. This new technology could improve the control of prosthetic hands and provide haptic feedback, more commonly known as the experience of touch, for amputees to reconnect a previously severed sense of touch.
“Significant research has been done on tactile sensors for artificial hands, but there is still a need for advances in lightweight, low-cost, robust multimodal tactile sensors,” said , Ph.D., senior author, an associate professor in the and a member of the Stiles-Nicholson Brain Institute, the Institute for Sensing and Embedded Network Systems Engineering (I-SENSE), and the Center for Complex Systems and Brain Sciences (), who conducted the study with first author and Ph.D. student Moaed A. Abd. “The tactile information from all the individual fingertips in our study provided the foundation for a higher hand-level of perception enabling the distinction between ten complex, multi-textured surfaces that would not have been possible using purely local information from an individual fingertip. We believe that these tactile details could be useful in the future to afford a more realistic experience for prosthetic hand users through an advanced haptic display, which could enrich the amputee-prosthesis interface and prevent amputees from abandoning their prosthetic hand.”
Researchers compared four different machine learning algorithms for their successful classification capabilities: K-nearest neighbor (KNN), support vector machine (SVM), random forest (RF), and neural network (NN). The time-frequency features of the liquid metal sensors were extracted to train and test the machine learning algorithms. The NN generally performed the best at the speed and texture detection with a single finger and had a 99.2 percent accuracy to distinguish between ten different multi-textured surfaces using four liquid metal sensors from four fingers simultaneously.
“The loss of an upper limb can be a daunting challenge for an individual who is trying to seamlessly engage in regular activities,” said , Ph.D., dean, College of Engineering and Computer Science. “Although advances in prosthetic limbs have been beneficial and allow amputees to better perform their daily duties, they do not provide them with sensory information such as touch. They also don’t enable them to control the prosthetic limb naturally with their minds. With this latest technology from our research team, we are one step closer to providing people all over the world with a more natural prosthetic device that can ‘feel’ and respond to its environment.”
Study co-authors are Rudy Paul, Department of Ocean and Mechanical Engineering; Aparna Aravelli, Ph.D.; Ou Bai, Ph.D.; and Leonel Lagos, Ph.D., PMP, all with ; and Maohua Lin, Ph.D., Department of Ocean and Mechanical Engineering.
The research was supported by the National Institute of Biomedical Imaging and Bioengineering of the National Institutes of Health (NIH) and the National Institute of Aging of the NIH, the National Science Foundation, the Department of Energy and pilot grants from the Stiles-Nicholson Brain Institute and I-SENSE.
--
Tags: students | engineering | faculty and staff | research | jupiter | technology