Leveraging Multiple Query Logs to Improve Language Models for Spoken Query Recognition

Xiao Li, Patrick Nguyen, Geoffrey Zweig, and Dan Bohus

Abstract

A voice search system requires a speech interface that can correctly recognize spoken queries uttered by users. The recognition performance strongly relies on a robust language model. In this work, we present the use of multiple data sources, with the focus on query logs, in improving ASR language models for a voice search application. Our contributions are three folds: (1) the use of text queries from web search and mobile search in language modeling; (2) the use of web click data to predict query forms from business listing forms; and (3) the use of voice query logs in creating a positive feedback loop. Experiments show that by leveraging these resources, we can achieve recognition performance comparable to, or even better than, that of a previously deploy system where a large amount of spoken query transcripts are used in language modeling.

Details

Publication typeInproceedings
Published inICASSP
PublisherIEEE
> Publications > Leveraging Multiple Query Logs to Improve Language Models for Spoken Query Recognition