Maximum Expected BLEU Training of Phrase and Lexicon Translation Models

This paper proposes a new discriminative training method in constructing phrase and lexicon translation models. In order to reliably learn a myriad of parameters in these models, we propose a smoothed BLEU score-based utility function with KL regularization as the objective, and train the models on a large parallel dataset. For training, we derive growth transformations for phrase and lexicon translation probabilities to iteratively improve the objective. The proposed method, evaluated on the Europarl German-to-English dataset, leads to a 1.1 BLEU point improvement over a state-of-the-art baseline translation system. In IWSLT 2011 Benchmark, our system using the proposed method achieves the best Chinese-to-English translation result on the task of translating TED talks.

HeDeng_acl12_dtmt_camera_07.pdf
PDF file

In  Proceedings of ACL

Publisher  Association for Computational Linguistics

Details

TypeArticle
> Publications > Maximum Expected BLEU Training of Phrase and Lexicon Translation Models