In a breakthrough study, which researchers peg as a crucial finding in understanding how brain processes spoken language, researchers at Duke and Massachusetts Institute of Technology have identified the area of the brain that is sensitive to the timing of speech.
As anyone would agree, timing and rhythm are important to the structure of human speech and according to researchers our auditory system likely takes shortcuts to deal with the onslaught of information. Researchers say that it may be done by sampling information in chunks similar in length to that of an average consonant or syllable.
For the research, study co-author Tobias Overath, an assistant research professor of psychology and neuroscience at Duke and corresponding author Josh McDermott from MIT, cut recordings of foreign speech into short chunks ranging from 30 to 960 milliseconds in length, and then reassembled the pieces using a novel algorithm to create new sounds that the authors call ‘speech quilts’. The shorter the pieces of the resulting speech quilts, the greater the disruption was to the original structure of the speech.
These speech quilts were then played to participants and at the same time activity of neurons was measured in real time using functional magnetic resonance imaging machine. The team hypothesized that brain areas involved in speech processing would show larger responses to speech quilts made up of longer segments.
Their hypothesis turned out to be correct as a region of the brain called the superior temporal sulcus (STS) became highly active during the 480- and 960-millisecond quilts compared with the 30-millisecond quilts. In contrast, other areas of the brain involved in processing sound did not change their response as a result of the differences in the sound quilts. The superior temporal sulcus is known to integrate auditory and other sensory information. But no one has shown that the STS is sensitive to time structures in speech.
“That was pretty exciting. We knew we were onto something,” said Overath, who is a member of the Duke Institute for Brain Sciences.
In a bid to ensure that their results are accurate and to rule out any other possible explanations for the activation of the STS, the team tested numerous control sounds they created to mimic speech. One of the synthetic sounds they created shared the frequency of speech but lacked its rhythms. Another removed all the pitch from the speech. A third used environmental sounds.
They quilted each of these control stimuli, chopping them up in either 30- or 960-millisecond pieces and stitching them back together, before playing them to participants. The STS didn’t seem responsive to the quilting manipulation when it was applied to these control sounds.
“We really went to great lengths to be certain that the effect we were seeing in STS was due to speech-specific processing and not due to some other explanation, for example, pitch in the sound or it being a natural sound as opposed to some computer-generated sound,” Overath said.
The group plans to study whether the response in STS is similar for foreign speech that is phonetically much different than English, such as Mandarin, or quilts of familiar speech that is intelligible and has meaning. For familiar speech they might see stronger activation on the left side of the brain, which is thought to be dominant in processing language.
The study has been published in journal Nature Neuroscience.