脳が多数の単語を処理する際の仕組みを解明(When is the Brain Like a Subway Station? When It’s Processing Many Words at Once)

ad

2025-10-20 ニューヨーク大学(NYU)

ニューヨーク大学(NYU)の研究チームは、人間の脳が複数の単語を同時に理解する仕組みを明らかにし、その処理を「地下鉄駅」に例えて説明した。研究では脳波と神経画像データを解析し、脳が単語を順に処理するのではなく、複数の単語を重なり合う時間軸で並行的に処理していることを発見。音声として入る言葉が瞬時に音響的特徴、語彙、意味、文脈の順で階層的に分析され、それぞれの処理が入れ替わりながら進行することが確認された。これは、列車が駅に到着・停車・出発を繰り返すように、脳が複数の情報流を効率的に整理していることを示す。研究成果は『Proceedings of the National Academy of Sciences(PNAS)』に掲載され、言語障害の理解やAI言語モデルの設計への応用が期待されている。

脳が多数の単語を処理する際の仕組みを解明(When is the Brain Like a Subway Station? When It’s Processing Many Words at Once)
The figure shows how the brain works to decode the different aspects of words over time, with phonetics (i.e., sounds) processed most quickly and semantic meaning coming taking longer. Because people are hearing these words in the context of a story, the meanings of words can be anticipated, resulting in semantic meaning being processed earlier than phonetics. Image courtesy of Laura Gwilliams.

<関連情報>

階層的動的符号化が人間の脳内の音声理解を調整する Hierarchical dynamic coding coordinates speech comprehension in the human brain

Laura Gwilliams, Alec Marantz, David Poeppel, and Jean-Rémi King
Proceedings of the National Academy of Sciences  Published:October 17, 2025
DOI:https://doi.org/10.1073/pnas.2422097122

Significance

To understand speech, the brain generates a hierarchy of neural representations, which map from sound to meaning. We recorded whole-brain activity while participants listened to audiobooks and modeled neural activity using time-resolved machine learning methods. Our analyses reveal that different neural patterns are activated over time to process each feature of the hierarchy, and abstract features have a slower spatial trajectory than sensory features. This spatiotemporally dynamic code enables a local history of inputs to be encoded in parallel, across a hierarchy from sound to meaning, also in parallel. This leads us to propose a dynamic model of cortical language processing: hierarchical dynamic coding (HDC).

Abstract

Speech comprehension involves transforming an acoustic waveform into meaning. To do so, the human brain generates a hierarchy of features that converts the sensory input into increasingly abstract language properties. However, little is known about how rapid incoming sequences of hierarchical features are continuously coordinated. Here, we propose that each language feature is supported by a dynamic neural code, which represents the sequence history of hierarchical features in parallel. To test this “hierarchical dynamic coding” (HDC) hypothesis, we use time-resolved decoding of brain activity to track the construction, maintenance, and update of a comprehensive hierarchy of language features spanning phonetic, word form, lexical–syntactic, syntactic, and semantic representations. For this, we recorded 21 native English participants with magnetoencephalography (MEG), while they listened to two hours of short stories in English. Our analyses reveal three main findings. First, the brain represents and simultaneously maintains a sequence of hierarchical features. Second, the duration of these representations depends on their level in the language hierarchy. Third, each representation is maintained by a dynamic neural code, which evolves at a speed commensurate with its corresponding linguistic level. This HDC preserves the maintenance of information over time while limiting destructive interference between successive features. Overall, HDC reveals how the human brain maintains and updates the continuously unfolding language hierarchy during natural speech comprehension, thereby anchoring linguistic theories to their biological implementations.

医療・健康
ad
ad
Follow
ad
タイトルとURLをコピーしました