skip to main content
10.1145/2487575.2487630acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
poster

Direct optimization of ranking measures for learning to rank models

Published:11 August 2013Publication History

ABSTRACT

We present a novel learning algorithm, DirectRank, which directly and exactly optimizes ranking measures without resorting to any upper bounds or approximations. Our approach is essentially an iterative coordinate ascent method. In each iteration, we choose one coordinate and only update the corresponding parameter, with all others remaining fixed. Since the ranking measure is a stepwise function of a single parameter, we propose a novel line search algorithm that can locate the interval with the best ranking measure along this coordinate quite efficiently. In order to stabilize our system in small datasets, we construct a probabilistic framework for document-query pairs to maximize the likelihood of the objective permutation of top-$\tau$ documents. This iterative procedure ensures convergence. Furthermore, we integrate regression trees as our weak learners in order to consider the correlation between the different features. Experiments on LETOR datasets and two large datasets, Yahoo challenge data and Microsoft 30K web data, show an improvement over state-of-the-art systems.

References

  1. M. Bazaraa, H. Sherali and C. Shetty. Nonlinear Programming: Theory and Algorithms, 3rd Edition. Wiley-Interscience, 2006.Google ScholarGoogle ScholarCross RefCross Ref
  2. D. Bertsimas, A. Chang and C. Rudin. Integer Optimization Methods for Supervised Ranking. Technical Report, 388--11, 2011.Google ScholarGoogle Scholar
  3. S. Boyd and L. Vandenberghe. Convex Optimization. Cambridge University, 2004. Google ScholarGoogle ScholarCross RefCross Ref
  4. C. Burges, R. Ragno and Q. Le. Learning to Rank with Nonsmooth Cost Functions. NIPS, 19:193--200, 2007.Google ScholarGoogle Scholar
  5. C. Burges, T. Shaked, E. Renshaw, A. Lazier, M. Deeds, N. Hamilton, and G. Hullender. Learning to Rank Using Gradient Descent. ICML, 89--96, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. C. Burges. From RankNet to LambdaRank to LambdaMART: An Overview. Microsoft Research Technical Report, MSR-TR-2010-82.Google ScholarGoogle Scholar
  7. C. Calauzenes, N. Usunier, P. Gallinari. On the (Non-)existence of Convex, Calibrated Surrogate Losses for Ranking. NIPS, 197--205, 2012Google ScholarGoogle Scholar
  8. Z. Cao, T. Qin, T. Liu, M. Tsai and H. Li. Learning to Rank: From Pairwise to Listwise Approach. ICML, 129--136, 2007 Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. S. Chakrabarti, R. Khanna, U. Sawant, and C. Bhattacharyya. Structured Learning for Non-smooth Ranking Losses. KDD, 88--96, 2008 Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. O. Chapelle and Y. Chang. Yahoo! Learning to Rank Challenge Overview. JMLR-Proceedings Track, 14:1--24, 2011Google ScholarGoogle Scholar
  11. O. Chapelle and M. Wu. Gradient Descent Optimization of Smoothed Information Retrieval Metrics. Information Retrieval, 13(3):216--235, 2010 Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. W. Cohen, R. Schapire, and Y. Singer. Learning to Order Things. JAIR, 10:243--270, 1999. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. K. Crammer and Y. Singer. Pranking with Ranking. NIPS, 14:641--647, 2001.Google ScholarGoogle Scholar
  14. Y. Freund, R. Iyer, R. Schapire and Y. Singer. An Efficient Boosting Algorithm for Combining Preferences. JMLR, 4:933--969, 2003. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. J. Friedman. Greedy Function Approximation: A Gradient Boosting Machine. The Annals of Statistics, 29(5):1189--1232, 2001.Google ScholarGoogle ScholarCross RefCross Ref
  16. E. Harrington. Online Ranking/Collaborative Filtering Using the Perceptron Algorithm. ICML Workshop then Conference, 20(1):250, 2003.Google ScholarGoogle Scholar
  17. R. Herbrich, T. Graepel, and K. Obermayer. Support Vector Learning for Ordinal Regression. ICANN, 97--102, 1999.Google ScholarGoogle Scholar
  18. T. Joachims. Optimizing Search Engines Using Clickthrough Data. KDD, 1:7--102, 2002.Google ScholarGoogle Scholar
  19. J. Kuo, P. Cheng, and H. Wang. Learning to Rank from Bayesian Decision Inference. CIKM, 827--836, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Q. Le and A. Smola. Direct Optimization of Ranking Measures. NICTA Tech report, 2007.Google ScholarGoogle Scholar
  21. T. Liu. Learning to Rank for Information Retrieval. Springer, 2011.Google ScholarGoogle ScholarCross RefCross Ref
  22. D. McAllester, T. Hazan, and J. Keshet. Direct loss minimization for structured prediction. ICML, 23:1594--1602, 2010.Google ScholarGoogle Scholar
  23. D. Metzler and W. Croft. Linear Feature-Based Models for Information Retrieval. Information Retrieval, 10(3):257--274, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. F. Och. Minimum Error Rate Training in Statistical Machine Translation. ACL, 311--318, 2003.Google ScholarGoogle Scholar
  25. R. Plackett. The Analysis of Permutations. Applied Statistics, 193--202, 1975.Google ScholarGoogle Scholar
  26. T. Qin, T. Liu, and H. Li. A General Approximation Framework for Direct Optimization of Information Retrieval Measures, Information Retrieval, 13(4):375--397, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. T. Qin, T. Liu, J. Xu, and H. Li. LETOR: A Benchmark Collection for Research on Learning to Rank for Information Retrieval. Information Retrieval, 13(4):346--374, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. T. Qin, X. Zhang, M. Tsai, D. Wang, T. Liu, and H. Li. Query-level Loss Functions for Information Retrieval. Information Processing and Management, 44(2):838--855, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. P. Ravikumar, A. Tewari, and E. Yang. On NDCG Consistency of Listwise Ranking Methods. AISTATS, 618--626, 2011.Google ScholarGoogle Scholar
  30. C. Rudin. The P-Norm Push: A Simple Convex Ranking Algorithm that Concentrates at the Top of the List. JMLR, 10:2233--2271, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. R. Schapire and Y. Freund. Boosting: Foundations and Algorithms. MIT Press, 2012. Google ScholarGoogle ScholarCross RefCross Ref
  32. A. Shashua and A. Levin. Ranking with Large Margin Principle: Two Approaches. NIPS, 937--944, 2003.Google ScholarGoogle Scholar
  33. M. Tan, T. Xia, L. Guo and S. Wang. Direct Optimization of Ranking Measures for Learning to Rank Models. Technical Report, 2013.Google ScholarGoogle Scholar
  34. M. Taylor, J. Guiver, S. Robertson and T. Minka. Softrank: Optimizing Non-smooth Rank Metrics. WSDM, 77--86, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  35. M. Tsai, T. Liu, Q. Tao, H. Chen and W. Ma. Frank: A Ranking Method with Fidelity Loss. SIGIR, 383--390, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. S. Tyree, K. Weinberger, K. Agrawal, and J. Paykin. Parallel boosted regression trees for web search ranking. WWW, 387--396, 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. H. Valizadegan, R. Jin, R. Zhang, and J. Mao. Learning to Rank by Optimizing NDCG Measure. NIPS, 1883--1891, 2009.Google ScholarGoogle Scholar
  38. V. Vapnik. Statistical Learning Theory. John Wiley, 1998.Google ScholarGoogle ScholarDigital LibraryDigital Library
  39. F. Xia, T. Liu, J. Wang, W. Zhang, and H. Li. Listwise Approach to Learning to Rank: Theory and Algorithm. ICML, 1192--1199, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  40. J. Xu, T. Liu, M. Lu, H. Li, and W. Ma. Directly Optimizing Evaluation Measures in Learning to Rank. SIGIR, 107--114, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  41. J. Xu and H. Li. AdaRank: A Boosting Algorithm for Information Retrieval. SIGIR, 391--398, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  42. Y. Yue, T. Finley, F. Radlinski, and T. Joachims. A Support Vector Method for Optimizing Average Precision. SIGIR, 271--278, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  43. Q. Wu, C. Burges, K. Svore, and J. Gao. Adapting Boosting for Information Retrieval Measures. Information Retrieval, 13(3):254--270, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Direct optimization of ranking measures for learning to rank models

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      KDD '13: Proceedings of the 19th ACM SIGKDD international conference on Knowledge discovery and data mining
      August 2013
      1534 pages
      ISBN:9781450321747
      DOI:10.1145/2487575

      Copyright © 2013 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 11 August 2013

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • poster

      Acceptance Rates

      KDD '13 Paper Acceptance Rate125of726submissions,17%Overall Acceptance Rate1,133of8,635submissions,13%

      Upcoming Conference

      KDD '24

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader