Can you hear what I say? New findings on human speech recognition

Can you hear what I say? – New findings on human speech recognition
Visualization of the medial geniculate body (MGB) in the brain of human test persons. The MGB is part of the auditory pathway. It is organized in such a way that certain sections represent certain sound frequencies. The study shows that the part of the MGB that transmits information from the ear to the cerebral cortex is actively involved in human speech recognition. Credit: Dresden University of Technology

Neuroscientists at TU Dresden were able to prove that speech recognition in humans begins in the sensory pathways from the ear to the cerebral cortex and not, as previously assumed, exclusively in the cerebral cortex itself.

In many households, it is impossible to imagine life without language assistants—they switch devices on or off, report on news from all over the world or know what the weather will be like tomorrow. The speech of these systems is mostly based on , a branch of artificial intelligence. The machine generates its knowledge from recurring patterns of data. In recent years, the use of artificial neural networks has largely improved computer-based speech recognition.

For neuroscientist Professor Katharina von Kriegstein from TU Dresden, however, the remains the "most admirable speech processing machine." "It works much better than computer-based speech processing and will probably continue to do so for a long time to come," comments Professor von Kriegstein, "because the exact processes of speech processing in the brain are still largely unknown."

In a recent study, the neuroscientist from Dresden and her team discovered another building block in the mystery of human speech processing. In the study, 33 test persons were examined using functional magnetic resonance imaging (MRI). The test persons received speech signals from different speakers. They were asked to perform a speech task or a control task for in random order. The team of scientists recorded the brain activity of the test persons during the experiment using MRI. The evaluation of the recordings showed that a structure in the left auditory pathway—the ventral medial geniculate body, (vMGB) - has particularly high activity when the test persons perform a speech task (in contrast to the control task) and when the test persons are particularly good at recognizing speech.

Previously, it was assumed that all auditory information was equally transmitted via the auditory pathways from the ear to the cerebral cortex. The current recordings of the increased activity of the vMGB show that the processing of the auditory information begins before the auditory pathways reach the cerebral cortex. Katharina von Kriegstein explains the results as follows: "For some time now, we have had the first indications that the auditory pathways are more specialized in speech than previously assumed. This study shows that this is indeed the case: The part of the vMGB that transports information from the ear to the cerebral cortex processes auditory information differently when speech is to be recognized than when other components of communication signals are to be recognized, such as the speaker's voice for example."

The recognition of auditory speech is of extreme importance for interpersonal communication. Understanding the underlying neuronal processes will not only be important for the further development of computer-based .

These new results may also have relevance for some of the symptoms of developmental dyslexia. It is known that the left MGB functions differently in dyslexic persons than in others. A specialization of the left MGB in speech may explain why dyslexic people often have difficulty understanding signals in noisy environments (such as restaurants). Katharina von Kriegstein and her team are now going to carry out further studies in order to scientifically prove these indications.

More information: Paul Glad Mihai et al. Modulation of tonotopic ventral medial geniculate body is behaviorally relevant for speech recognition, eLife (2019). DOI: 10.7554/eLife.44837

Journal information: eLife
Citation: Can you hear what I say? New findings on human speech recognition (2019, August 28) retrieved 20 April 2024 from https://medicalxpress.com/news/2019-08-human-speech-recognition.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Neuroscientists discover neural mechanisms of developmental dyslexia

48 shares

Feedback to editors