Rescoring a phrase-based machine transliteration system with recurrent neural network language models

  • Authors:
  • Andrew Finch;Paul Dixon;Eiichiro Sumita

  • Affiliations:
  • NICT, Hikaridai Keihanna Science City, Japan;NICT, Hikaridai Keihanna Science City, Japan;NICT, Hikaridai Keihanna Science City, Japan

  • Venue:
  • NEWS '12 Proceedings of the 4th Named Entity Workshop
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

The system entered into this year's shared transliteration evaluation is implemented within a phrase-based statistical machine transliteration (SMT) framework. The system is based on a joint source-channel model in combination with a target language model and models to control the length of the sequences generated. The joint source-channel model was trained using a many-to-many Bayesian bilingual alignment. The focus of this year's system is on input representation. In order attempt to mitigate data sparseness issues in the joint source-channel model, we augmented the system with recurrent neural network (RNN) models that can learn to project the grapheme set onto a smaller hidden representation. We performed experiments on development data to evaluate the effectiveness of our approach. Our results show that using an RNN language model can improve performance for language pairs with large grapheme sets on the target side.