Electrocorticography reveals continuous auditory and visual speech tracking in temporal and occipital cortex

Eur J Neurosci. 2020 Mar;51(5):1364-1376. doi: 10.1111/ejn.13992. Epub 2018 Aug 12.

Abstract

During natural speech perception, humans must parse temporally continuous auditory and visual speech signals into sequences of words. However, most studies of speech perception present only single words or syllables. We used electrocorticography (subdural electrodes implanted on the brains of epileptic patients) to investigate the neural mechanisms for processing continuous audiovisual speech signals consisting of individual sentences. Using partial correlation analysis, we found that posterior superior temporal gyrus (pSTG) and medial occipital cortex tracked both the auditory and the visual speech envelopes. These same regions, as well as inferior temporal cortex, responded more strongly to a dynamic video of a talking face compared to auditory speech paired with a static face. Occipital cortex and pSTG carry temporal information about both auditory and visual speech dynamics. Visual speech tracking in pSTG may be a mechanism for enhancing perception of degraded auditory speech.

Keywords: audiovisual speech; clear speech; continuous speech; multisensory; naturalistic stimuli.

Publication types

  • Research Support, N.I.H., Extramural
  • Research Support, Non-U.S. Gov't
  • Research Support, U.S. Gov't, Non-P.H.S.

MeSH terms

  • Acoustic Stimulation
  • Auditory Cortex*
  • Auditory Perception
  • Brain Mapping
  • Electrocorticography
  • Humans
  • Occipital Lobe
  • Speech
  • Speech Perception*
  • Visual Perception