Learning Term-weighting Functions for Similarity Measures

Measuring the similarity between two texts is a fundamental problem in many NLP and IR applications. Among the existing approaches, the cosine measure of the term vectors representing the original texts has been widely used, where the score of each term is often determined by a TFIDF formula. Despite its simplicity, the quality of such cosine similarity measure is usually domain dependent and decided by the choice of the term-weighting function. In this paper, we propose a novel framework that learns the term-weighting function. Given the labeled pairs of texts as training data, the learning procedure tunes the model parameters by minimizing the specified loss function of the similarity score. Compared to traditional TFIDF term-weighting schemes, our approach shows a significant improvement on tasks such as judging the quality of query suggestions and filtering irrelevant ads for online advertising.

Yih-EMNLP-09.pdf
PDF file

In  Conference on Empirical Methods in Natural Language Processing

Publisher  Association for Computational Linguistics
All copyrights reserved by ACL 2009

Details

TypeInproceedings
> Publications > Learning Term-weighting Functions for Similarity Measures