Preference-based learning to rank.

Machine Learning (Impact Factor: 1.69). 09/2010; 80:189-211. DOI: 10.1007/s10994-010-5176-9
Source: DBLP

ABSTRACT This paper presents an efficient preference-based ranking algorithm running in two stages. In the first stage, the algorithm
learns a preference function defined over pairs, as in a standard binary classification problem. In the second stage, it makes
use of that preference function to produce an accurate ranking, thereby reducing the learning problem of ranking to binary
classification. This reduction is based on the familiar QuickSort and guarantees an expected pairwise misranking loss of at
most twice that of the binary classifier derived in the first stage. Furthermore, in the important special case of bipartite
ranking, the factor of two in loss is reduced to one. This improved bound also applies to the regret achieved by our ranking
and that of the binary classifier obtained.

Our algorithm is randomized, but we prove a lower bound for any deterministic reduction of ranking to binary classification
showing that randomization is necessary to achieve our guarantees. This, and a recent result by Balcan et al., who show a
regret bound of two for a deterministic algorithm in the bipartite case, suggest a trade-off between achieving low regret
and determinism in this context.

Our reduction also admits an improved running time guarantee with respect to that deterministic algorithm. In particular,
the number of calls to the preference function in the reduction is improved from Ω(n
2) to O(nlog n). In addition, when the top k ranked elements only are required (k≪n), as in many applications in information extraction or search engine design, the time complexity of our algorithm can be
further reduced to O(klog k+n). Our algorithm is thus practical for realistic applications where the number of points to rank exceeds several thousand.

  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: We provide a summary of the mathematical and computational techniques that have enabled learning reductions to effectively address a wide class of problems, and show that this approach to solving machine learning problems can be broadly useful.
  • [Show abstract] [Hide abstract]
    ABSTRACT: The recent years have witnessed the birth and explosive growth of the web. It is obvious that the exponential growth of the web has made it into a huge interconnected source of information wherein finding a document without a searching tool is unimaginable. Today's search engines try to provide the most relevant suggestions to the user queries. To do this, different strategies are used to enhance the precision of the information retrieval process. In this paper, a learning method is proposed to rank the web documents in a search engine. The proposed method takes advantage of the user feedback to enhance the precision of the search results. To do so, it uses a learning automata-based approach to train the search engine. In this method, the user feedback is defined as its interest to review an item. Within the search results, the document that is visited by the user is more likely relevant to the user query. Therefore, its choice probability must be increased by the learning automaton. By this, the rank of the most relevant documents increases as that of the others decreases. To investigate the efficiency of the proposed method, extensive simulation experiment is conducted on well-known data collections. The obtained results show the superiority of the proposed approach over the existing methods in terms of mean average precision, precision at position n, and normalized discount cumulative gain.
    Decision Support Systems 12/2012; 54(1):574–583. DOI:10.1016/j.dss.2012.08.005 · 2.04 Impact Factor
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: Ranking is one of the most fundamental problems in machine learning with applications in many branches of computer science such as: information retrieval systems, recommendation systems, machine translation and computational biology. Ranking objects based on possibly conflicting preferences is a central problem in voting research and social choice theory. In this paper we present a new simple combinatorial ranking algorithm adapted to the preference-based setting. We apply this new algorithm to the well-known scenario where the edges of the preference tournament are determined by the majority-voting model. It outperforms existing methods when it cannot be assumed that there exists global ranking of good enough quality and applies combinatorial techniques that havent been used in the ranking context before. Performed experiments show the superiority of the new algorithm over existing methods, also over these that were designed to handle heavily perturbed statistics. By combining our techniques with those presented in \cite{mohri}, we obtain a purely combinatorial algorithm that answers correctly most of the queries in the heterogeneous scenario, where the preference tournament is only locally of good quality but is not necessarily pseudotransitive. As a byproduct of our methods, we obtain the algorithm solving clustering problem for the directed planted partition model. To the best of our knowledge, it is the first purely combinatorial algorithm tackling this problem.


Available from