Speech emotional features extraction based on electroglottograph

  • Authors:
  • Lijiang Chen;Xia Mao;Pengfei Wei;Angelo Compare

  • Affiliations:
  • -;-;-;-

  • Venue:
  • Neural Computation
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

This study proposes two classes of speech emotional features extracted from electroglottography EGG and speech signal. The power-law distribution coefficients PLDC of voiced segments duration, pitch rise duration, and pitch down duration are obtained to reflect the information of vocal folds excitation. The real discrete cosine transform coefficients of the normalized spectrum of EGG and speech signal are calculated to reflect the information of vocal tract modulation. Two experiments are carried out. One is of proposed features and traditional features based on sequential forward floating search and sequential backward floating search. The other is the comparative emotion recognition based on support vector machine. The results show that proposed features are better than those commonly used in the case of speaker-independent and content-independent speech emotion recognition.