Decoding hidden brain conversations to advance neuroprostheses
Reconstructing the stable information from neurons driving skills like tying sneakers can aid paralyzed patients
- Created: January 09, 2020
- Published Version
Once you learn a skill like tying your sneakers, you can perform it consistently over the years. This suggests the neural activity in the brain associated with the skill remains stable over time.
Thus far, though, it has been impossible to find this stable neural signal, because the electrodes necessary to monitor the brain activity see an ever-changing set of neurons.
New Northwestern Medicine findings provide an immediate way of compensating for this change in recorded neurons. Hidden in the activity patterns, scientists discovered a small set of neural "conversations” underlying the individual neurons' activity that described movements in a remarkably stable way over many months or years.
Northwestern scientists were able to record, decode and reconstruct preserved activity patterns for common movement skills from these ever-changing neurons in both motor and sensory areas of the cerebral cortex.
The findings have immediate implications for neuroprostheses, devices that bypass neurological injuries by inferring intended motor actions from a patient’s brain, then using this decoded information to provide the patient voluntary control of a computer cursor, a robot limb or even their own reanimated limb.
“The study is significant because it is the first to extract or ‘decode’ a stable pattern of information from many neurons across the different brain areas that work together to create a movement,” said lead author Lee Miller, professor of physiology at Northwestern University Feinberg School of Medicine. “Previous studies focused on the individual neurons, whose activity cannot typically be recorded for more than a few hours, and reported somewhat contradictory findings.”
Sara Solla, also a lead author and professor of physiology at Feinberg, emphasized the importance of considering the concerted activity of many neurons. “It is not the activity of single neurons that allows us to understand how the brain works, but the collective activity of neurons interconnected to form a network,” she said.
The challenge the scientists solved is a bit like attending a cocktail party where many people are talking to each other about politics, religion and the weather. The combined voices are impossible to follow, and listening to a single person only works until they wander away. However, by monitoring a few representative people and extracting the information related to these three different topics, one could come away from the evening with a very good running commentary of current events.
Such is the case in the cerebral cortex: one can sample several hundred neurons out of many million to find the dominant neural "conversations." By combining the simultaneous recordings from many neurons, it is possible to extract stable patterns of activity distributed across the whole population, potentially allowing for consistent control of a neuroprosthesis over long periods of time.
The study was published Jan. 6 in Nature Neuroscience.
Although there have been pre-clinical demonstrations of neuroprostheses that try to address multiple movement commands, they all have the same limitation, Miller said. The individual neurons driving the neuroprostheses change across days. This instability in the signals being recorded makes it difficult to interpret movement commands, and requires the patient and the neuroprosthesis to relearn how to move almost every day.
Other groups have developed methods to accelerate this process, but the Northwestern results suggest that it may not be necessary if one focuses on the neural conversations rather than on the individual neurons.
“Our study supports a recent model of how the brain works: by using highly distributed patterns of activity across neurons to generate behavior, rather than by carefully adjusting the activity of each individual neuron,” said first author Juan Gallego, a former postdoc in Miller’s lab who is now an assistant professor at Imperial College London.
Miller’s final goal is to have a neuroprosthesis that translates brain signals (the activity of tens to hundreds of neurons) into a variety of movements over a long period of time. Current neuroprostheses only function particularly well for the tasks on which they have been trained.
“By accurately predicting how the subject wants to activate their many individual muscles, this future neuroprosthesis should, in principle, allow the user to perform any movement she wants,” Miller said.
But challenges remain to achieve this. Scientists can currently build a mathematical model that predicts muscle activity during a given task (and Miller’s team has done it for several tasks), or even during a slightly different task.
“Yet, these mathematical models fail when you test them on a very different task,” Miller said. “In other words, with current methods a ‘hammering nails’ model may not work well when the subject is ‘applying lipstick.’”
Miller’s lab team is working to overcome this challenge. In preclinical studies, they are performing wireless recordings of brain signals, muscle activity and movement data. They are exploring methods from artificial intelligence to build a model that works for a much broader variety of movements than has previously been possible.
“If successful, this research will open new exciting avenues for bringing neuroprostheses from the lab to the patients’ homes,” Miller said.
Coauthors on the paper were Matt Perich at Icahn School of Medicine at Mount Sinai and Raeed Chowdhury at University of Pittsburgh.
The work was supported in part by grants F31-NS092356 and NS053603 from the National Institute of Neurological Disorder and Stroke and grant T32-HD07418 from the National Institute of Child Health and Human Development, all of the National Institutes of Health, the National Science Foundation.