skip to main content
10.1145/2396761.2398550acmconferencesArticle/Chapter ViewAbstractPublication PagescikmConference Proceedingsconference-collections
short-paper

A comprehensive analysis of parameter settings for novelty-biased cumulative gain

Authors Info & Claims
Published:29 October 2012Publication History

ABSTRACT

In the TREC Web Diversity track, novelty-biased cumulative gain (α-NDCG) is one of the official measures to assess retrieval performance of IR systems. The measure is characterised by a parameter, α, the effect of which has not been thoroughly investigated. We find that common settings of α, i.e. α=0.5, may prevent the measure from behaving as desired when evaluating result diversification. This is because it excessively penalises systems that cover many intents while it rewards those that redundantly cover only few intents. This issue is crucial since it highly influences systems at top ranks. We revisit our previously proposed threshold, suggesting α be set on a query-basis. The intuitiveness of the measure is then studied by examining actual rankings from TREC 09-10 Web track submissions. By varying α according to our query-based threshold, the discriminative power of α-NDCG is not harmed and in fact, our approach improves α-NDCG's robustness. Experimental results show that the threshold for α can turn the measure to be more intuitive than using its common settings.

References

  1. B. Carterette. System effectiveness, user models, and user utility: a conceptual framework for investigation. In SIGIR '11, pages 903--912, 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. O. Chapelle, D. Metlzer, Y. Zhang, and P. Grinspan. Expected reciprocal rank for graded relevance. In CIKM '09, pages 621--630, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. C. L. Clarke, N. Craswell, and I. Soboroff. Overview of the trec 2009 web track. In TREC 18, 2009.Google ScholarGoogle Scholar
  4. C. L. Clarke, M. Kolla, G. V. Cormack, O. Vechtomova, A. Ashkan, S. Büttcher, and I. MacKinnon. Novelty and diversity in information retrieval evaluation. In SIGIR '08, pages 659--666, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. C. L. Clarke, M. Kolla, and O. Vechtomova. An effectiveness measure for ambiguous and underspecified queries. In ICTIR '09, pages 188--199, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. T. Leelanupab, G. Zuccon, and J. M. Jose. A query-basis approach to parametrizing novelty-biased cumulative gain. In ICTIR '11, pages 327--331, 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. T. Leelanupab. A Ranking Framework and Evaluation for Diversity-Based Retrieval. PhD thesis, School of Computing Science, University of Glasgow, 2012.Google ScholarGoogle Scholar
  8. F. Radlinski, P. N. Bennett, B. Carterette, and T. Joachims. Redundancy, diversity and interdependent document relevance. ACM SIGIR Forum, 43:46--52, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. S. E. Robertson, M. Maron, and W. Cooper. Probability of Relevance: A Unification of two Competing Models for Document Retrieval. Information technology: research and development, pages 1--21, 1982.Google ScholarGoogle Scholar
  10. T. Sakai. Evaluating evaluation metrics based on the bootstrap. In SIGIR '06, pages 525--532, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. T. Sakai, N. Craswell, R. Song, S. Robertson, Z. Dou, and C. Y. Lin. Simple evaluation metrics for diversified search results. In EVIA '10, pages 42--50, 2010.Google ScholarGoogle Scholar
  12. T. Sakai and R. Song. Evaluating diversified search results using per-intent graded relevance. In SIGIR '11, pages 1043--1052, Beijing, China, 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. E. Yilmaz, J. Aslam, and S. Robertson. A new rank correlation coefficient for information retrieval. In SIGIR '08, pages 587--594, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. C. X. Zhai, W. W. Cohen, and J. Lafferty. Beyond independent relevance: methods and evaluation metrics for subtopic retrieval. In SIGIR '03, pages 10--17, 2003. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. G. Zuccon. Ranking Documents with Quantum Probabilities. PhD thesis, School of Computing Science, University of Glasgow, 2012.Google ScholarGoogle Scholar

Index Terms

  1. A comprehensive analysis of parameter settings for novelty-biased cumulative gain

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      CIKM '12: Proceedings of the 21st ACM international conference on Information and knowledge management
      October 2012
      2840 pages
      ISBN:9781450311564
      DOI:10.1145/2396761

      Copyright © 2012 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 29 October 2012

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • short-paper

      Acceptance Rates

      Overall Acceptance Rate1,861of8,427submissions,22%

      Upcoming Conference

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader