A new study has uncovered how the brain seamlessly transforms sounds, speech patterns, and words into the flow of everyday conversations. Using advanced technology to analyze over 100 hours of brain activity during real-life discussions, researchers revealed the intricate pathways that allow us to effortlessly speak and understand. These insights not only deepen our understanding of human connection but also pave the way for transformative advancements in speech technology and communication tools.
A new study led by Dr. Ariel Goldstein, from the Department of Cognitive and Brain Sciences and the Business School at the Hebrew University of Jerusalem, Google Research, in collaboration with the Hasson Lab at the Neuroscience Institute at Princeton University, Dr. Flinker and Dr. Devinsky from the NYU Langone Comprehensive Epilepsy Center, has developed a unified computational framework to explore the neural basis of human conversations. This research bridges acoustic, speech, and word-level linguistic structures, offering unprecedented insights into how the brain processes everyday speech in real-world settings.
The study, published in Nature Human Behaviour, recorded brain activity over 100 hours of natural, open-ended conversations using a technique called electrocorticography (ECoG). To analyze this data, the team used a speech-to-text model called Whisper, which helps break down language into three levels: simple sounds, speech patterns, and the meaning of words. These layers were then compared to brain activity using advanced computer models.
The results showed that the framework could predict brain activity with great accuracy. Even when applied to conversations that were not part of the original data, the model correctly matched different parts of the brain to specific language functions. For example, regions involved in hearing and speaking aligned with sound and speech patterns, while areas involved in higher-level understanding aligned with the meanings of words.
The study also found that the brain processes language in a sequence. Before we speak, our brain moves from thinking about words to forming sounds, while after we listen, it works backward to make sense of what was said. The framework used in this study was more effective than older methods at capturing these complex processes.
Our findings help us understand how the brain processes conversations in real-life settings. By connecting different layers of language, we're uncovering the mechanics behind something we all do naturally-talking and understanding each other."
Dr. Ariel Goldstein, from the Department of Cognitive and Brain Sciences and the Business School at the Hebrew University of Jerusalem
This research has potential practical applications, from improving speech recognition technology to developing better tools for people with communication challenges. It also offers new insights into how the brain makes conversation feel so effortless, whether it's chatting with a friend or engaging in a debate. The study marks an important step toward building more advanced tools to study how the brain handles language in real-world situations.
Source:
Journal reference:
Goldstein, A., et al. (2025). A unified acoustic-to-speech-to-language embedding space captures the neural basis of natural language processing in everyday conversations. Nature Human Behaviour. doi.org/10.1038/s41562-025-02105-9.