skip to main content
10.1145/3331184.3331347acmconferencesArticle/Chapter ViewAbstractPublication PagesirConference Proceedingsconference-collections
short-paper
Open Access

Revisiting Approximate Metric Optimization in the Age of Deep Neural Networks

Published:18 July 2019Publication History

ABSTRACT

Learning-to-Rank is a branch of supervised machine learning that seeks to produce an ordering of a list of items such that the utility of the ranked list is maximized. Unlike most machine learning techniques, however, the objective cannot be directly optimized using gradient descent methods as it is either discontinuous or flat everywhere. As such, learning-to-rank methods often optimize a loss function that either is loosely related to or upper-bounds a ranking utility instead. A notable exception is the approximation framework originally proposed by Qin et al. that facilitates a more direct approach to ranking metric optimization. We revisit that framework almost a decade later in light of recent advances in neural networks and demonstrate its superiority empirically. Through this study, we hope to show that the ideas from that work are more relevant than ever and can lay the foundation of learning-to-rank research in the age of deep neural networks.

References

  1. Mart'in Abadi, Paul Barham, Jianmin Chen, Zhifeng Chen, Andy Davis, Jeffrey Dean, Matthieu Devin, Sanjay Ghemawat, Geoffrey Irving, Michael Isard, et almbox. 2016. Tensorflow: a system for large-scale machine learning. In Proc. of the 12th USENIX Symposium on Operating Systems Design and Implementation. 265--283. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Chris Burges, Tal Shaked, Erin Renshaw, Ari Lazier, Matt Deeds, Nicole Hamilton, and Greg Hullender. 2005. Learning to rank using gradient descent. In Proc. of the 22nd International Conference on Machine Learning. 89--96. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Christopher J.C. Burges. 2010. From RankNet to LambdaRank to LambdaMART: An Overview. Technical Report Technical Report MSR-TR-2010--82. Microsoft Research.Google ScholarGoogle Scholar
  4. Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li. 2007. Learning to rank: from pairwise approach to listwise approach. In Proc. of the 24th International Conference on Machine Learning. 129--136. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Olivier Chapelle and Yi Chang. 2011. Yahoo! learning to rank challenge overview. In Proc. of the Learning to Rank Challenge. 1--24. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Jerome H Friedman. 2001. Greedy function approximation: a gradient boosting machine. Annals of Statistics, Vol. 29, 5 (2001), 1189--1232.Google ScholarGoogle ScholarCross RefCross Ref
  7. Sergey Ioffe and Christian Szegedy. 2015. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proc. of the 32nd International Conference on Machine Learning (ICML). 448--456. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Kalervo J"arvelin and Jaana Kekalainen. 2002. Cumulated gain-based evaluation of IR techniques. ACM Transactions on Information Systems, Vol. 20, 4 (2002), 422--446. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Thorsten Joachims. 2006. Training linear SVMs in linear time. In Proc. of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 217--226. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Guolin Ke, Qi Meng, Thomas Finley, Taifeng Wang, Wei Chen, Weidong Ma, Qiwei Ye, and Tie-Yan Liu. 2017. LightGBM: A Highly Efficient Gradient Boosting Decision Tree. In Advances in Neural Information Processing Systems 30. 3146--3154. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Donald A Metzler, W Bruce Croft, and Andrew Mccallum. 2005. Direct maximization of rank-based metrics for information retrieval. CIIR report 429. University of Massachusetts.Google ScholarGoogle Scholar
  12. Rama Kumar Pasumarthi, Sebastian Bruch, Xuanhui Wang, Cheng Li, Michael Bendersky, Marc Najork, Jan Pfeifer, Nadav Golbandi, Rohan Anil, and Stephan Wolf. 2019. TF-Ranking: Scalable TensorFlow Library for Learning-to-Rank. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. (to appear). Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Tao Qin and Tie-Yan Liu. 2013. Introducing LETOR 4.0 Datasets. (2013). arxiv: 1306.2597Google ScholarGoogle Scholar
  14. Tao Qin, Tie-Yan Liu, and Hang Li. 2010. A general approximation framework for direct optimization of information retrieval measures. Information Retrieval, Vol. 13, 4 (2010), 375--397. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Michael Taylor, John Guiver, Stephen Robertson, and Tom Minka. 2008. SoftRank: Optimizing Non-smooth Rank Metrics. In Proc. of the 1st International Conference on Web Search and Data Mining. 77--86. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Xuanhui Wang, Cheng Li, Nadav Golbandi, Michael Bendersky, and Marc Najork. 2018. The LambdaLoss Framework for Ranking Metric Optimization. In Proc. of the 27th ACM International Conference on Information and Knowledge Management. 1313--1322. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Qiang Wu, Christopher JC Burges, Krysta M Svore, and Jianfeng Gao. 2010. Adapting boosting for information retrieval measures. Information Retrieval, Vol. 13, 3 (2010), 254--270. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Fen Xia, Tie-Yan Liu, Jue Wang, Wensheng Zhang, and Hang Li. 2008. Listwise approach to learning to rank: theory and algorithm. In Proc. of the 25th International Conference on Machine Learning. 1192--1199. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Jun Xu and Hang Li. 2007. AdaRank: A Boosting Algorithm for Information Retrieval. In Proc. of the 30th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval. 391--398. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Revisiting Approximate Metric Optimization in the Age of Deep Neural Networks

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      SIGIR'19: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval
      July 2019
      1512 pages
      ISBN:9781450361729
      DOI:10.1145/3331184

      Copyright © 2019 Owner/Author

      Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 18 July 2019

      Check for updates

      Qualifiers

      • short-paper

      Acceptance Rates

      SIGIR'19 Paper Acceptance Rate84of426submissions,20%Overall Acceptance Rate792of3,983submissions,20%

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader