Scalable HMM based inference engine in large vocabulary continuous speech recognition

  • Authors:
  • Jike Chong;Kisun You;Youngmin Yi;Ekaterina Gonina;Christopher Hughes;Wonyong Sung;Kurt Keutzer

  • Affiliations:
  • Department of Electrical Engineering and Computer Science, University of California, Berkeley;School of Electrical Engineering, Seoul National University and Intel Corporation;Department of Electrical Engineering and Computer Science, University of California, Berkeley;Department of Electrical Engineering and Computer Science, University of California, Berkeley;Intel Corporation;School of Electrical Engineering, Seoul National University;Department of Electrical Engineering and Computer Science, University of California, Berkeley

  • Venue:
  • ICME'09 Proceedings of the 2009 IEEE international conference on Multimedia and Expo
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

Parallel scalability allows an application to efficiently utilize an increasing number of processing elements. In this paper we explore a design space for application scalability for an inference engine in large vocabulary continuous speech recognition (LVCSR). Our implementation of the inference engine involves a parallel graph traversal through an irregular graph-based knowledge network with millions of states and arcs. The challenge is not only to define a software architecture that exposes sufficient fine-grained application concurrency, but also to efficiently synchronize between an increasing number of concurrent tasks and to effectively utilize the parallelism opportunities in today's highly parallel processors. We propose four application-level implementation alternatives we call "algorithm styles", and construct highly optimized implementations on two parallel platforms: an Intel Core i7 multicore processor and a NVIDIA GTX280 manycore processor. The highest performing algorithm style varies with the implementation platform. On 44 minutes of speech data set, we demonstrate substantial speedups of 3.4× on Core i7 and 10.5× on GTX280 compared to a highly optimized sequential implementation on Core i7 without sacrificing accuracy. The parallel implementations contain less than 2.5% sequential overhead, promising scalability and significant potential for further speedup on future platforms.