skip to main content
10.1145/1571941.1572021acmconferencesArticle/Chapter ViewAbstractPublication PagesirConference Proceedingsconference-collections
research-article

On the local optimality of LambdaRank

Published:19 July 2009Publication History

ABSTRACT

A machine learning approach to learning to rank trains a model to optimize a target evaluation measure with repect to training data. Currently, existing information retrieval measures are impossible to optimize directly except for models with a very small number of parameters. The IR community thus faces a major challenge: how to optimize IR measures of interest directly. In this paper, we present a solution. Specifically, we show that LambdaRank, which smoothly approximates the gradient of the target measure, can be adapted to work with four popular IR target evaluation measures using the same underlying gradient construction. It is likely, therefore, that this construction is extendable to other evaluation measures. We empirically show that LambdaRank finds a locally optimal solution for mean NDCG@10, mean NDCG, MAP and MRR with a 99% confidence rate. We also show that the amount of effective training data varies with IR measure and that with a sufficiently large training set size, matching the training optimization measure to the target evaluation measure yields the best accuracy.

References

  1. C.J.C. Burges, R. Ragno, and Q.V. Le. Learning to rank with nonsmooth cost functions. In Neural Information Processing Systems (NIPS), 2006. See also MSR Technical Report MSR-TR-2006-60.Google ScholarGoogle Scholar
  2. C.J.C. Burges, T. Shaked, E. Renshaw, A. Lazier, M. Deeds, N. Hamilton, and G. Hullender. Learning to rank using gradient descent. In International Conference on Machine Learning (ICML), Bonn, Germany, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Z. Cao, T. Qin, T.Y. Liu, M.F. Tsai, and H. Li. Learning to rank: From pairwise to listwise approach. In International Conference on Machine Learning (ICML), pages 129--136, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. K. Crammer and Y. Singer. Pranking with ranking. In Neural Information Processing Systems (NIPS), 2001.Google ScholarGoogle Scholar
  5. R. Herbrich, T. Graepel, and K. Obermayer. Large margin rank boundaries for ordinal regression. Advances in Large Margin Classifiers, pages 115--132, 2000.Google ScholarGoogle Scholar
  6. T. Qin, T.Y. Liu, and H. Li. A general approximation framework for direct optimization of information retrieval measures. Microsoft Technical Report MSR-TR-2008-164, 2008.Google ScholarGoogle Scholar
  7. T. Qin, X.-D. Zhang, M.-F. Tsai, D.-S. Wang, T.-Y. Liu, and H. Li. Query-level loss functions for information retrieval. Information Processing and Management, 44(2):838--855, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. S. Robertson and H. Zaragoza. On rank-based effectiveness measures and optimization. Information Processing and Management, 10:321--339, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. B. Taskar, V. Chatalbashev, D. Koller, and C. Guestrin. Learning structured prediction models: A large margin approach. In International Conference on Machine Learning (ICML), Bonn, Germany, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. I. Tsochantaridis, T. Hofmann, T. Joachims, and Y. Altun. Support vector machine learning for interdependent and structured output spaces. In International Conference on Machine Learning (ICML), 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. J. Xu and H. Li. Adarank: A boosting algorithm for information retrieval. In ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR), pages 391--398, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Y. Yue and C.J.C Burges. On using simultaneous perturbation stochastic approximation for ir measures, and the empirical optimality of lambdarank. NIPS Machine Learning for Web Search Workshop, 2007.Google ScholarGoogle Scholar
  13. Y. Yue, T. Finley, F. Radlinski, and T. Joachims. A support vector method for optimizing average precision. In ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR), 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. On the local optimality of LambdaRank

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Conferences
        SIGIR '09: Proceedings of the 32nd international ACM SIGIR conference on Research and development in information retrieval
        July 2009
        896 pages
        ISBN:9781605584836
        DOI:10.1145/1571941

        Copyright © 2009 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 19 July 2009

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article

        Acceptance Rates

        Overall Acceptance Rate792of3,983submissions,20%

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader