A medical first- Doctors enable quadriplegic man control prosthetic arms with thoughts
Researchers at the Applied Physics Laboratory of Johns Hopkins University and Johns Hopkins Medicine have enabled a quadriplegic patient to control two robotic arms with his brain – and even feed himself after two years of research.The technology uses a system of devices that automates a portion of the robotic control with artificial intelligence.
The team had implanted six electrode arrays into both sides of the brain of Robert "Buz" Chmielewski, a quadriplegic patient with minimal movement and feeling in his hands and fingers as part of a clinical trial. With the help of them, the researchers planned to teach a patient Chmielewski to use two advanced prosthetic arms – and not only to control them but also to feel what the robotic arms are touching.
Researchers demonstrated simultaneous control of two of the prosthetic limbs through a brain-machine interface developed by APL. Very little is known about how exactly the interface works, but usually, such devices work on the basis of an algorithm that learns to read and recognize the activity that appears in response to the presentation of a certain action and then use it to control the robotic arm.
The patient managed to learn how to move the robotic arms within a year, and now he has already managed to master more precise actions – to cope with a knife and fork. The smart software puts his utensils in roughly the right spot, and then Buz uses his brain signals to cut the food with a knife and fork. Once he is done cutting, the software then brings the food near his mouth, where he again uses brain signals to bring the food the last several inches to his mouth so he can eat it.
Buz's accomplishment marks a big step toward restoring function and autonomy for patients affected by an illness or injury that results in the partial or total loss of use of all four limbs and torso.
"It's pretty cool," said Chmielewski, whose sense of accomplishment was unmistakable after using his thoughts to command the robotic limbs to cut and feed him a piece of golden sponge cake. "I wanted to be able to do more of it," he said.
Nearly two years ago, Chmielewski underwent a 10-hour brain surgery at the Johns Hopkins Hospital in Baltimore as part of a clinical trial — originally spearheaded by the Defense Advanced Research Projects Agency, and leveraging advanced prosthetic limbs developed by the Johns Hopkins Applied Physics Laboratory (APL) in Laurel, Maryland — to allow participants to control assistive devices, and enable perception of physical stimuli (touching the limbs) using neurosignals from the brain.
Surgeons implanted six electrode arrays into both sides of his brain, and within months he was able to demonstrate, for the first time, simultaneous control of two of the prosthetic limbs through a brain-machine interface developed by APL.
Researchers were impressed with his progress during the first year of testing and wanted to further push the bounds of what could be accomplished. Using an internal research grant from APL, the team launched a parallel line of inquiry — termed "Smart Prosthetics" — to develop strategies for providing advanced robot control and sensory feedback from both hands at the same time using neural stimulation. That team included Francesco Tenore, David Handelman, Andrew Badger, Matthew Fifer and Luke Osborn from APL, as well as Tessy Thomas, Robert Nickl, Nathan Crone, Gabriela Cantarero and Pablo Celnik from the School of Medicine.
They set out to develop a closed-loop system that merges artificial intelligence, robotics and a brain-machine interface. In the instance of Chmielewski serving himself dessert, the system enabled him to control the movements necessary to cut food with a fork and knife and feed himself.
"Our ultimate goal is to make activities such as eating easy to accomplish, having the robot do one part of the work and leaving the user, in this case Buz, in charge of the details: which food to eat, where to cut, how big the cut piece should be," explained Handelman, an APL senior roboticist specializing in human-machine teaming. "By combining brain-computer interface signals with robotics and artificial intelligence, we allow the human to focus on the parts of the task that matter most."
Tenore, an APL neuroscientist and principal investigator for the Smart Prosthetics study, said the next steps for this effort include not only expanding the number and types of activities of daily living that Buz can demonstrate with this form of human-machine collaboration, "but also providing him with additional sensory feedback as he's conducting these tasks, so that he won't have to rely entirely on vision to know if he's succeeding."
"The idea is that he'd experience this the same way that uninjured people can 'feel' how they're tying their shoelaces, for example, without having to look at what they're doing," Tenore said.
In an interview just before Thanksgiving — the traditional launch of a food-heavy holiday season — Buz reflected on the significance of this research for individuals with limited mobility. Disabilities like his take away a person's independence, he said, particularly their ability to eat by themselves.
"A lot of people take that for granted," he said. "To be able to do this independently and still be able to interact with family is a game-changer."