Auditory cortex encodes lipreading information through spatially distributed activity
Recommended Citation
Karthik G, Cao CZ, Demidenko MI, Jahn A, Stacey WC, Wasade VS, and Brang D. Auditory cortex encodes lipreading information through spatially distributed activity. Curr Biol 2024.
Document Type
Article
Publication Date
9-9-2024
Publication Title
Current biology
Abstract
Watching a speaker's face improves speech perception accuracy. This benefit is enabled, in part, by implicit lipreading abilities present in the general population. While it is established that lipreading can alter the perception of a heard word, it is unknown how these visual signals are represented in the auditory system or how they interact with auditory speech representations. One influential, but untested, hypothesis is that visual speech modulates the population-coded representations of phonetic and phonemic features in the auditory system. This model is largely supported by data showing that silent lipreading evokes activity in the auditory cortex, but these activations could alternatively reflect general effects of arousal or attention or the encoding of non-linguistic features such as visual timing information. This gap limits our understanding of how vision supports speech perception. To test the hypothesis that the auditory system encodes visual speech information, we acquired functional magnetic resonance imaging (fMRI) data from healthy adults and intracranial recordings from electrodes implanted in patients with epilepsy during auditory and visual speech perception tasks. Across both datasets, linear classifiers successfully decoded the identity of silently lipread words using the spatial pattern of auditory cortex responses. Examining the time course of classification using intracranial recordings, lipread words were classified at earlier time points relative to heard words, suggesting a predictive mechanism for facilitating speech. These results support a model in which the auditory system combines the joint neural distributions evoked by heard and lipread words to generate a more precise estimate of what was said.
Medical Subject Headings
Humans; Auditory Cortex; Speech Perception; Adult; Magnetic Resonance Imaging; Female; Male; Lipreading; Young Adult; Visual Perception
PubMed ID
39153482
ePublication
ePub ahead of print
Volume
34
Issue
17
First Page
4021
Last Page
4032