Study discovers how brain processes, stores words that we hear
According to neuroscientists from Georgetown University Medical Centre, the brain's auditory lexicon, or catalogue of verbal language, is located in the front of the primary auditory cortex, not the rear - a discovery that challenges a century-long understanding of this part of the brain.
WASHINGTON: According to neuroscientists from Georgetown University Medical Centre, the brain's auditory lexicon, or catalogue of verbal language, is located in the front of the primary auditory cortex, not the rear - a discovery that challenges a century-long understanding of this part of the brain.
The new knowledge is significant because it may have an influence on healing and rehabilitation following brain damage such as a stroke. The findings of the study were published in Neurobiology of Language.
Riesenhuber's group discovered a lexicon for written words near the base of the left hemisphere of the brain, known as the Visual Word Form Area (VWFA), and later discovered that newly learnt written words are added to the VWFA. The current study intended to see if a comparable vocabulary for spoken words exists in the Auditory Word Form Area (AWFA), which is situated anterior to the main auditory cortex.
"Since the early 1900s, scientists believed spoken word recognition took place behind the primary auditory cortex, but that model did not fit well with many observations from patients with speech recognition deficits, such as stroke patients," says Maximilian Riesenhuber, PhD, professor in the Department of Neuroscience at Georgetown University Medical Center and senior author of this study.
"Our discovery of an auditory lexicon more towards the front of the brain provides a new target area to help us understand speech comprehension deficits." In the study, led by Srikanth Damera, MD, PhD, 26 volunteers went through three rounds of functional magnetic resonance imaging (fMRI) scans to examine their spoken word processing abilities.
The technique used in this study was called functional-MRI rapid adaptation (fMRI-RA), which is more sensitive than conventional fMRI in assessing representation of auditory words as well as the learning of new words.
"In future studies, it will be interesting to investigate how interventions directed at the AWFA affect speech comprehension deficits in populations with different types of strokes or brain injury," says Riesenhuber.
"We are also trying to understand how the written and spoken word systems interact. Beyond that, we are using the same techniques to look for auditory lexica in other parts of the brain, such as those responsible for speech production."
Josef Rauschecker, PhD, DSc, professor in the Department of Neuroscience at Georgetown and co-author of the study, adds that many aspects of how the brain processes words, either written or verbal, remain unexplored.
"We know that when we learn to speak, we rely on our auditory system to tell us whether the sound we've produced accurately represents our intended word," he said. "We use that feedback to refine future attempts to say the word.
However, the brain's process for this remains poorly understood - both for young children learning to speak for the first time, but also for older people learning a second language."