An emerging research area named Learning-to-Rank (LtR) has shown that effective solutions to the ranking problem can leverage machine learning techniques applied to a large set of features capturing the relevance of a candidate document for the user query. Large-scale search systems must however answer user queries very fast, and the computation of the features for candidate documents must comply with strict back-end latency constraints. The number of features cannot thus grow beyond a given limit, and Feature Selection (FS) techniques have to be exploited to find a subset of features that both meets latency requirements and leads to high effectiveness of the trained models. In this paper, we propose three new algorithms for FS specifically designed for the LtR context where hundreds of continuous or categorical features can be involved. We present a comprehensive experimental analysis conducted on publicly available LtR datasets and we show that the proposed strategies outperform a well-known state-of-theart competitor.
|Data di pubblicazione:||2016|
|Titolo:||Fast feature selection for learning to rank|
|Titolo del libro:||ICTIR 2016 - Proceedings of the 2016 ACM International Conference on the Theory of Information Retrieval|
|Digital Object Identifier (DOI):||http://dx.doi.org/10.1145/2970398.2970433|
|Appare nelle tipologie:||4.1 Articolo in Atti di convegno|