Cortical processing of temporal modulations

  • Authors:
  • Xiaoqin Wang;Thomas Lu;Li Liang

  • Affiliations:
  • Laboratory of Auditory Neurophysiology, Department of Biomedical Engineering. Johns Hopkins University School of Medicine, 720 Rutland Avenue, Ross 424, Baltimore, MD;Laboratory of Auditory Neurophysiology, Department of Biomedical Engineering. Johns Hopkins University School of Medicine, 720 Rutland Avenue, Ross 424, Baltimore, MD;Laboratory of Auditory Neurophysiology, Department of Biomedical Engineering. Johns Hopkins Univ. School of Medicine, Baltimore, MD and Hearing Center, Pear River Hospital of First Medical Univ., ...

  • Venue:
  • Speech Communication - Special issue on the nature of speech perception (the psychophysics of speech perception III)
  • Year:
  • 2003

Quantified Score

Hi-index 0.00

Visualization

Abstract

Temporal modulations are fundamental components of human speech and animal communication sounds. Understanding their representations in the auditory cortex is a crucial step towards our understanding of brain mechanisms underlying speech processing. While modulated signals have long been used as experimental stimuli, their cortical representations are not completely understood, particularly for rapid modulations. Known physiological data do not adequately explain psychophysical observations on the perception of rapid modulations, largely due to slow stimulus-synchronized temporal discharge patterns of cortical neurons. In this article, we summarize recent findings from our laboratory on temporal processing mechanisms in the auditory cortex. These findings show that the auditory cortex represents slow modulations explicitly using a temporal code and fast modulations implicitly by a discharge rate code. Rapidly modulated signals within a short-time window (∼20-30 ms) are integrated and transformed into a discharge rate-based representation. The findings also indicate that there is a shared representation of temporal modulations by cortical neurons that encodes the temporal profile embedded in complex sounds of various spectral contents. Our results suggest that cortical processing of sound streams operates on a "segment-by-segment" basis with a temporal integration window on the order of ∼20-30 ms.