Simultaneous Synchronization of Text and Speech for Broadcast News Subtitling

  • Authors:
  • Jie Gao;Qingwei Zhao;Ta Li;Yonghong Yan

  • Affiliations:
  • ThinkIT Speech Lab, Institute of Acoustics, Chinese Academy of Sciences, Beijing, P.R. China 100190;ThinkIT Speech Lab, Institute of Acoustics, Chinese Academy of Sciences, Beijing, P.R. China 100190;ThinkIT Speech Lab, Institute of Acoustics, Chinese Academy of Sciences, Beijing, P.R. China 100190;ThinkIT Speech Lab, Institute of Acoustics, Chinese Academy of Sciences, Beijing, P.R. China 100190

  • Venue:
  • ISNN 2009 Proceedings of the 6th International Symposium on Neural Networks: Advances in Neural Networks - Part III
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, we present our initial effort in automatic generation of subtitle for live broadcast news programs, utilizing the fact that nearly perfect transcriptions are available. Instead of using the former error-prone automatic-speech-recognition (ASR)-based method, we propose to formulate the subtitling problem as synchronization of text and speech, which is further simplified into an anchor points estimation problem. The Viterbi algorithm for hidden Markov model (HMM) is augmented with new criterions for the online anchor points estimation. Experiments indicate that our proposed methods show satisfying performance for the simultaneous subtitling application. We also present a brief introduction into our whole subtitling system under further development.