Researchers use brain-computer interfaces technology to decode the brain’s mechanism

When people suffer debilitating injuries or illnesses of the nervous system, they sometimes lose the ability to perform tasks normally taken for granted, such as walking, playing music or driving a car. They can imagine doing something, but the injury might block that action from occurring.

Brain-computer interface systems exist that can translate brain signals into a desired action to regain some function, but they can be a burden to use because they don't always operate smoothly and need readjustment to complete even simple tasks.

Researchers at the University of Pittsburgh and Carnegie Mellon University are working on understanding how the brain works when learning tasks with the help of brain-computer interface technology.

In a set of papers, the second of which was published today in Nature Biomedical Engineering, the team is moving the needle forward on brain-computer interface technology intended to help improve the lives of amputee patients who use neural prosthetics.

Let's say during your work day, you plan out your evening trip to the grocery store. That plan is maintained somewhere in your brain throughout the day, but probably doesn't reach your motor cortex until you actually get to the store. We're developing brain-computer interface technologies that will hopefully one day function at the level of our everyday intentions."

Aaron Batista, Associate Professor of Bioengineering, Swanson School of Engineering, University of Pittsburgh

Batista, Pitt postdoctoral research associate Emily Oby and the Carnegie Mellon researchers have collaborated on developing direct pathways from the brain to external devices.

They use electrodes smaller than a hair that record neural activity and make it available for control algorithms.

In the team's first study, published last June in the Proceedings of the National Academy of Sciences, the group examined how the brain changes with the learning of new brain-computer interface skills.

"When the subjects form a motor intention, it causes patterns of activity across those electrodes, and we render those as movements on a computer screen.

The subjects then alter their neural activity patterns in a manner that evokes the movements that they want," said project co-director Steven Chase, a professor of biomedical engineering at the Neuroscience Institute at Carnegie Mellon.

In the new study, the team designed technology whereby the brain-computer interface readjusts itself continually in the background to ensure the system is always in calibration and ready to use.

"We change how the neural activity affects the movement of the cursor, and this evokes learning," said Pitt's Oby, the study's lead author. "If we changed that relationship in a certain way, it required that our animal subjects produce new patterns of neural activity to learn to control the movement of the cursor again. Doing so took them weeks of practice, and we could watch how the brain changed as they learned."

In a sense, the algorithm "learns" how to adjust to the noise and instability that is inherent in neural recording interfaces.

The findings suggest that the process for humans to master a new skill involves the generation of new neural activity patterns. The team eventually would like this technology to be used in a clinical setting for stroke rehabilitation.

Such self-recalibration procedures have been a long-sought goal in the field of neural prosthetics, and the method presented in the team's studies is able to recover automatically from instabilities without requiring the user to pause to recalibrate the system by themselves.

"Let's say that the instability was so large such that the subject was no longer able to control the brain-computer interface," said Yu. "Existing self-recalibration procedures are likely to struggle in that scenario, whereas in our method, we've demonstrated it can in many cases recover from even the most dramatic instabilities."

Source:
Journal reference:

Degenhart, A. D., et al. (2020) Stabilization of a brain–computer interface via the alignment of low-dimensional spaces of neural activity. Nature Biomedical Engineering. doi.org/10.1038/s41551-020-0542-9.

Comments

The opinions expressed here are the views of the writer and do not necessarily reflect the views and opinions of News Medical.
Post a new comment
Post

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.

You might also like...
Camouflage detection boosts neural networks for brain tumor diagnosis