Context Dependent Phonetic String Edit Distance for Automatic Speech Recognition

Jasha Droppo and Alex Acero

Abstract

An automatic speech recognition system searches for the word transcription with the highest overall score for a given acoustic observation sequence. This overall score is typically a weighted combination of a language model score and an acoustic model score. We propose including a third score, which measures the similarity of the word transcription's pronunciation to the output of a less constrained phonetic recognizer. We show how this phonetic string edit distance can be learned from data, and that including context in the model is essential for good performance. We demonstrate improved accuracy on a business search task.

Details

Publication typeInproceedings
Published inICASSP
PublisherIEEE
> Publications > Context Dependent Phonetic String Edit Distance for Automatic Speech Recognition