Mechanisms of enhancing visual-speech recognition by prior auditory information

Publikation: Beitrag in FachzeitschriftForschungsartikelBeigetragenBegutachtung

Beitragende

  • Helen Blank - , Max-Planck-Institut für Kognitions- und Neurowissenschaften (Autor:in)
  • Katharina von Kriegstein - , Max-Planck-Institut für Kognitions- und Neurowissenschaften (Autor:in)

Abstract

Speech recognition from visual-only faces is difficult, but can be improved by prior information about what is said. Here, we investigated how the human brain uses prior information from auditory speech to improve visual-speech recognition. In a functional magnetic resonance imaging study, participants performed a visual-speech recognition task, indicating whether the word spoken in visual-only videos matched the preceding auditory-only speech, and a control task (face-identity recognition) containing exactly the same stimuli. We localized a visual-speech processing network by contrasting activity during visual-speech recognition with the control task. Within this network, the left posterior superior temporal sulcus (STS) showed increased activity and interacted with auditory-speech areas if prior information from auditory speech did not match the visual speech. This mismatch-related activity and the functional connectivity to auditory-speech areas were specific for speech, i.e., they were not present in the control task. The mismatch-related activity correlated positively with performance, indicating that posterior STS was behaviorally relevant for visual-speech recognition. In line with predictive coding frameworks, these findings suggest that prediction error signals are produced if visually presented speech does not match the prediction from preceding auditory speech, and that this mechanism plays a role in optimizing visual-speech recognition by prior information.

Details

OriginalspracheEnglisch
Seiten (von - bis)109-118
Seitenumfang10
FachzeitschriftNeuroImage
Jahrgang65
PublikationsstatusVeröffentlicht - 15 Jan. 2013
Peer-Review-StatusJa
Extern publiziertJa

Externe IDs

PubMed 23023154
ORCID /0000-0001-7989-5860/work/142244404

Schlagworte

Schlagwörter

  • FMRI, Lip-reading, Multisensory, Predictive coding, Speech reading

Bibliotheksschlagworte