Skip to main content
  • 4683 Accesses

Abstract

In this chapter, we take the official evaluation results published at the LETOR website as the source to perform discussions on the performances of different learning-to-rank methods.

Note that there have been several other empirical studies [9, 12] in the literature, based on LETOR and other datasets. The conclusions drawn from these studies are similar to what we will introduce in this chapter.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

References

  1. Baeza-Yates, R., Ribeiro-Neto, B.: Modern Information Retrieval. Addison-Wesley, Reading (1999)

    Google Scholar 

  2. Cao, Z., Qin, T., Liu, T.Y., Tsai, M.F., Li, H.: Learning to rank: from pairwise approach to listwise approach. In: Proceedings of the 24th International Conference on Machine Learning (ICML 2007), pp. 129–136 (2007)

    Chapter  Google Scholar 

  3. Chapelle, O.: Direct optimization for ranking. In: Keynote Speech, SIGIR 2009 Workshop on Learning to Rank for Information Retrieval (LR4IR 2009) (2009)

    Google Scholar 

  4. Freund, Y., Iyer, R., Schapire, R., Singer, Y.: An efficient boosting algorithm for combining preferences. Journal of Machine Learning Research 4, 933–969 (2003)

    MathSciNet  Google Scholar 

  5. Herbrich, R., Obermayer, K., Graepel, T.: Large margin rank boundaries for ordinal regression. In: Advances in Large Margin Classifiers, pp. 115–132 (2000)

    Google Scholar 

  6. Järvelin, K., Kekäläinen, J.: Cumulated gain-based evaluation of IR techniques. ACM Transactions on Information Systems 20(4), 422–446 (2002)

    Article  Google Scholar 

  7. Joachims, T.: Optimizing search engines using clickthrough data. In: Proceedings of the 8th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD 2002), pp. 133–142 (2002)

    Google Scholar 

  8. Tsai, M.F., Liu, T.Y., Qin, T., Chen, H.H., Ma, W.Y.: Frank: a ranking method with fidelity loss. In: Proceedings of the 30th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2007), pp. 383–390 (2007)

    Chapter  Google Scholar 

  9. Verberne, S., Halteren, H.V., Theijssen, D., Raaijmakers, S., Boves, L.: Learning to rank qa data. In: SIGIR 2009 Workshop on Learning to Rank for Information Retrieval (LR4IR 2009) (2009)

    Google Scholar 

  10. Xu, J., Li, H.: Adarank: a boosting algorithm for information retrieval. In: Proceedings of the 30th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2007), pp. 391–398 (2007)

    Chapter  Google Scholar 

  11. Yue, Y., Finley, T., Radlinski, F., Joachims, T.: A support vector method for optimizing average precision. In: Proceedings of the 30th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2007), pp. 271–278 (2007)

    Chapter  Google Scholar 

  12. Zhang, M., Kuang, D., Hua, G., Liu, Y., Ma, S.: Is learning to rank effective for web search. In: SIGIR 2009 Workshop on Learning to Rank for Information Retrieval (LR4IR 2009) (2009)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Tie-Yan Liu .

Rights and permissions

Reprints and permissions

Copyright information

© 2011 Springer-Verlag Berlin Heidelberg

About this chapter

Cite this chapter

Liu, TY. (2011). Experimental Results on LETOR. In: Learning to Rank for Information Retrieval. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-14267-3_11

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-14267-3_11

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-14266-6

  • Online ISBN: 978-3-642-14267-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics