Share on Facebook Tweet on Twitter Share on LinkedIn Share by email
Search Vox: Leveraging multimodal refinement and partial knowledge for mobile voice search

Tim Paek, Bo Thiesson, Y. C. Ju, and Bongshin Lee

Abstract

Internet usage on mobile devices continues to grow as users seek anytime, anywhere access to information. Because users frequently search for businesses, directory assistance has been the focus of many voice search applications utilizing speech as the primary input modality. Unfortunately, mobile settings often contain noise which degrades performance. As such, we present Search Vox, a mobile search interface that not only facilitates touch and text refinement whenever speech fails, but also allows users to assist the recognizer via text hints. Search Vox can also take advantage of any partial knowledge users may have about the business listing by letting them express their uncertainty in an intuitive way using verbal wildcards. In simulation experiments conducted on real voice search data, leveraging multimodal refinement resulted in a 28% relative reduction in error rate. Providing text hints along with the spoken utterance resulted in even greater relative reduction, with dramatic gains in recovery for each additional character.

Details

Publication typeInproceedings
Published inProceedings of User Interface Software and Technology (UIST)
Pages141-150
PublisherAssociation for Computing Machinery, Inc.
> Publications > Search Vox: Leveraging multimodal refinement and partial knowledge for mobile voice search