Mechanisms of enhancing visual-speech recognition by prior auditory information

Research output: Contribution to journalResearch articleContributedpeer-review

Contributors

  • Helen Blank - , Max Planck Institute for Human Cognitive and Brain Sciences (Author)
  • Katharina von Kriegstein - , Max Planck Institute for Human Cognitive and Brain Sciences (Author)

Abstract

Speech recognition from visual-only faces is difficult, but can be improved by prior information about what is said. Here, we investigated how the human brain uses prior information from auditory speech to improve visual-speech recognition. In a functional magnetic resonance imaging study, participants performed a visual-speech recognition task, indicating whether the word spoken in visual-only videos matched the preceding auditory-only speech, and a control task (face-identity recognition) containing exactly the same stimuli. We localized a visual-speech processing network by contrasting activity during visual-speech recognition with the control task. Within this network, the left posterior superior temporal sulcus (STS) showed increased activity and interacted with auditory-speech areas if prior information from auditory speech did not match the visual speech. This mismatch-related activity and the functional connectivity to auditory-speech areas were specific for speech, i.e., they were not present in the control task. The mismatch-related activity correlated positively with performance, indicating that posterior STS was behaviorally relevant for visual-speech recognition. In line with predictive coding frameworks, these findings suggest that prediction error signals are produced if visually presented speech does not match the prediction from preceding auditory speech, and that this mechanism plays a role in optimizing visual-speech recognition by prior information.

Details

Original languageEnglish
Pages (from-to)109-118
Number of pages10
JournalNeuroImage
Volume65
Publication statusPublished - 15 Jan 2013
Peer-reviewedYes
Externally publishedYes

External IDs

PubMed 23023154
ORCID /0000-0001-7989-5860/work/142244404

Keywords

ASJC Scopus subject areas

Keywords

  • FMRI, Lip-reading, Multisensory, Predictive coding, Speech reading

Library keywords