Significant advance in our understanding of how sound is encoded for transmission to the brain

Scientists at Carnegie Mellon University have discovered that our ears use the most efficient way to process the sounds we hear, from babbling brooks to wailing babies.

These results represent a significant advance in our understanding of how sound is encoded for transmission to the brain, according to the authors, whose work is published with an accompanying "News and Views" editorial in the Feb. 23 issue of Nature.

The research provides a new mathematical framework for understanding sound processing and suggests that our hearing is highly optimized in terms of signal coding--the process by which sounds are translated into information by our brains--for the range of sounds we experience. The same work also has far-reaching, long-term technological implications, such as providing a predictive model to vastly improve signal processing for better quality compressed digital audio files and designing brain-like codes for cochlear implants, which restore hearing to the deaf.

To achieve their results, the researchers took a radically different approach to analyzing how the brain processes sound signals. Abstracting from the neural code at the auditory nerve, they represented sound as a discrete set of time points, or a "spike code," in which acoustic components are represented only in terms of their temporal relationship with each other. That's because the intensity and basic frequency of a given feature are essentially "kernalized," or compressed mathematically, into a single spike. This is similar to a player piano roll that can reproduce any song by recording what note to press when the spike code encodes any natural sound in terms of the precise timings of the elemental acoustic features. Remarkably, when the researchers derived the optimal set of features for natural sounds, they corresponded exactly to the patterns observed by neurophysiologists in the auditory nerves.

"We've found that timing of just a sparse number of spikes actually encodes the whole range of nature sounds, including components of speech such as vowels and consonants, and natural environment sounds like footsteps in a forest or a flowing stream," said Michael Lewicki, associate professor of computer science at Carnegie Mellon and a member of the Center for the Neural Basis of Cognition (CNBC). "We found that the optimal code for natural sounds is the same as that for speech. Oddly enough, cats share our own optimal auditory code for the English language."

"Our work is the only research to date that efficiently processes auditory code as kernalized spikes," said Evan Smith, a graduate student in psychology at the CNBC.

Until now, scientists and engineers have relied on Fourier transformations--initially discovered 200 years ago--to separate and re-constitute parameters like frequency and intensity as part of traditional sound signal processing.

"Our new signal processing framework appears far more efficient, effective and concise in conveying a rich variety of natural sounds than anything else," Lewicki said.

The approach by Smith and Lewicki dissects sound based only on the timing of compressed "spikes" associated with vowels (like cat vocalizations), consonants (like rocks hitting one another) and sibilants (ambient noise).

To gather sounds for their research, the scientists traipsed through the woods and recorded cracking branches, crunching leaves and wind rustling through leaves before returning to the laboratory to de-code the information contained in this rich set of sounds. They also discovered what they consider the most "natural" sound: if they play back a random set of spikes, it sounds like running water.

"We're very excited about this work because we can give a simple theoretical account of the auditory code which predicts how we could optimize signal processing to one day allow for much more efficient data storage on everything from DVDs to iPods," Lewicki said.

"For instance, if we could use a cochlear implant to 'talk' to the auditory nerve in a more natural way via our discovered coding, then we could quite possibly design implants that would convey sounds to the brain that are much more intelligible," he said.

Comments

The opinions expressed here are the views of the writer and do not necessarily reflect the views and opinions of News Medical.
Post a new comment
Post

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.

You might also like...
Lower pulmonary gas exchange may be linked to impaired cognitive function in long COVID patients