UEA researchers make surprising discovery about how the brain controls our hands

Researchers at the University of East Anglia have made an astonishing discovery about how our brains control our hands.

They used MRI data to study which parts of the brain are used when we handle tools, such as a knives.

They read out the signal from certain brain regions and tried to distinguish when participants handled tools appropriately for use.

Humans have used tools for millions of years, but this research is the first to show that actions such grasping a knife by its handle for cutting are represented by brain areas that also represent images of human hands, our primary ‘tool’ for interacting with the world.

The research could pave the way for the development of next-generation neuroprosthetics - prosthetic limbs that tap into the brain's control center, and help rehabilitate people who have lost function in their limbs due to brain injury.

The study was led by UEA and carried out at the Norfolk and Norwich University Hospital.

The emergence of handheld tools marks the beginning of a major discontinuity between humans and our closest primate relatives and is considered a defining feature of our species. Our findings could shed light on the regions of the brain that specifically evolved in the humans. We knew that seeing images of tools activates a different region of the brain to when we see other kinds of objects, for example a chair.

Until now it was assumed that the brain segregates visual information in this way to optimize processing of actions associated with tools. But how the human brain controls our hands to correctly grasp 3D objects such as tools was not well understood. We wanted to test whether the human brain automatically processes 3D objects in terms of how we grasp them for use. And we particularly wanted to find out whether we could use signals from specific parts of the brain to distinguish whether people were handling tools correctly - for example grasping a knife by the handle rather than the blade.”

Dr Stephanie Rossit, Lead researcher, UEA’s School of Psychology

The team used an MRI scanner to collect brain imaging data while participants interacted with 3D objects.

Dr Rossit said: “This was really challenging because the space inside the scanner is really small and the participants need to stay really still.

“So we used a one-of-a-kind ‘real action’ set-up for presenting 3D tools and other objects.

“Our participants lay in the dark, on a custom-built bed with a revolving table mounted above their waist, so that we could show them 3D objects and they could grasp them.

“We designed and 3D-printed everyday kitchen tools from non-magnetic materials so that they would be safe in the MRI such as a plastic knife, pizza cutter and a spoon as well as another group of 3D-printed bars to represent items that were not tools, which we used as control objects.”

Dr Ethan Knights, who was a PhD student with Dr Rossit, coordinated the data collection and scanned the brains of 20 volunteers at the Norfolk and Norwich University Hospital. In the first session, participants were asked to grasp the 3D tool and 3D bars correctly or incorrectly using the bespoke ‘real action’ set-up.

The same participants returned to the scanner for a second session in which they simply looked at pictures of tools and hands.

We studied brain activity when participants viewed pictures of tools and hands to identify which parts of the brain where the brain hand picture is represented. We then used state of the art machine learning to see if we could predict whether people actually grasped a tool by its handle or not. This is really important because knowing not to grasp an object, like a knife, by its blade, is critical to successful tool-use.

Dr Fraser Smith, UEA’s School of Psychology

Dr Rossit said: “In contrast to what most scientists thought, we were able to predict whether a tool was grasped correctly from the signals of brain areas that respond to the sight of pictures of hands and not from visual areas that respond to the sight of pictures of tools.

“Importantly the signals from the visual hand areas could only be used to predict hand actions with tools but could not predict matched actions with the control 3D bar objects.

“This suggests that the visual hand areas are specially tuned for actions with tools.

“Our discovery changes our fundamental understanding of how the brain controls our hands and could have important implications for health and society.

“For example, it could help develop better devices or rehabilitation for people who have lost function in their limbs due to brain injury. And it could even allow people without limbs to control prosthetics with their minds.

“The potential for brain-driven interfaces and prosthetics is very exciting,” she added.

‘Hand-selective visual regions represent how to grasp 3D tools: brain decoding during real actions’ is published in the Journal of Neuroscience on May 10, 2021. The study was funded by the BIAL Foundation.

Source:
Journal reference:

Knights, E., et al. (2021) Hand-selective visual regions represent how to grasp 3D tools: brain decoding during real actions. Journal of Neuroscience. doi.org/10.1523/JNEUROSCI.0083-21.2021.

Comments

The opinions expressed here are the views of the writer and do not necessarily reflect the views and opinions of News Medical.
Post a new comment
Post

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.

You might also like...
New study unveils why glioblastoma becomes resistant to treatment