Abstract
Large-scale test collections drive improvement in search technology to help users find information in free text.
- Buckley, C. and Voorhees, E. Retrieval system evaluation. In TREC: Experiment and Evaluation in Information Retrieval, E. Voorhees and D. Harman, Eds. MIT Press, Cambridge, MA, 2005, 53--75.Google Scholar
- Cleverdon, C. The Cranfield tests on index language devices. In Readings in Information Retrieval, K. Spärck Jones and P. Willett, Eds., Morgan Kaufmann, San Francisco, 1997. Google ScholarDigital Library
- Salton, G., Ed. The SMART Retrieval System: Experiments in Automatic Document Processing. Prentice-Hall, Inc., Englewood Cliffs, NJ, 1971. Google ScholarDigital Library
- Spärck Jones, K., Ed. Information Retrieval Experiment. Butterworths, London, 1981.Google Scholar
- Spärck Jones, K. and van Rijsbergen, C. Report on the need for and provision of an `ideal' information retrieval test collection. British Library Research and Development Report 5266, Computer Laboratory, University of Cambridge, 1975.Google Scholar
- Voorhees, E. and Harman, D., Eds. TREC: Experiment and Evaluation in Information Retrieval. MIT Press, Cambridge, MA, 2005. Google ScholarDigital Library
Index Terms
- TREC: Continuing information retrieval's tradition of experimentation
Recommendations
The TREC files: the (ground) truth is out there
SIGIR '11: Proceedings of the 34th international ACM SIGIR conference on Research and development in Information RetrievalTraditional tools for information retrieval (IR) evaluation, such as TREC's trec_eval, have outdated command-line interfaces with many unused features, or 'switches', accumulated over the years. They are usually seen as cumbersome applications by new IR ...
Liberal relevance criteria of TREC -: counting on negligible documents?
SIGIR '02: Proceedings of the 25th annual international ACM SIGIR conference on Research and development in information retrievalMost test collections (like TREC and CLEF) for experimental research in information retrieval apply binary relevance assessments. This paper introduces a four-point relevance scale and reports the findings of a project in which TREC-7 and TREC-8 ...
Relevance judgments between TREC and Non-TREC assessors
SIGIR '08: Proceedings of the 31st annual international ACM SIGIR conference on Research and development in information retrievalThis paper investigates the agreement of relevance assessments between official TREC judgments and those generated from an interactive IR experiment. Results show that 63% of documents judged relevant by our users matched official TREC judgments. ...
Comments