skip to main content
10.1145/1168812.1168842acmconferencesArticle/Chapter ViewAbstractPublication PagesiteConference Proceedingsconference-collections
Article

An infrastructure for the evaluation and comparison of information retrieval systems

Published:19 October 2006Publication History

ABSTRACT

Even though information retrieval systems have been successfully deployed for over 45 years, the field continues to evolve in performance, functionality, and accuracy. There are hundreds of different products available with different indexing and retrieval characteristics. How does one choose the appropriate system for a given application? The first step in that choice is the creation of a framework for comparison of IR products and an infrastructure that supports automated execution and analysis of testing results. The next step is providing an environment for subjective measurement using human evaluators. In this paper we briefly introduce the concepts used in IR system evaluation and report on our initial implementation of a framework for evaluating indexing performance. We also report a test case, which provides a comparative analysis of the indexing characteristics for three IR system implementations using a common collection of documents.

References

  1. Baeza-Yates, R. and Ribeiro-Neto, B. Modern Information Retrieval. Addison-Wesley Longman, Reading, MA, 1999. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Brown, A. B. and Seltzer, M. I. Operating system benchmarking in the wake of lmbench: A case study of the performance of NetBSD on the Intel x86 architecture. In Proceedings of the 1997 ACM SIGMETRICS international conference on Measurement and modeling of computer systems (Seattle, Washington, United States, June 15-18, 1997) ACM Press, New York NY, 1997, 214--224. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Croft, W. B. and Turtle, H. R. Text Retrieval and Inference. In P. S. Jacobs (ed.) Text-Based Intelligent Systems: Current Research and Practice in Information Extraction and Retrieval. Lawrence Erlbaum Associates, Hillsdale, New Jersey, 1992, 127--155. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Fox, C. Lexical analysis and stoplists. In W. B. Frakes and R. Baeza-Yates (eds.) Information Retrieval: Data Structures and Algorithms. Prentice-Hall, Englewood Cliffs, NJ, 1992, 102--130. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Gao, X., Murugesan, S., and Lo, B. Multi-dimensional evaluation of information retrieval results. In Proceedings of the IEEE/WIC/ACM International Conference (WI '04) (September 20-24, 2004). IEEE Computer Society, Washington, DC, 2004, 192--198. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Hart, M. S. History and Philosophy of Project Gutenberg, 1992. Retrieved August 15, 2006, from http://www.gutenberg.org/about/historyGoogle ScholarGoogle Scholar
  7. Salton, G., Wong, A., and Yang, C. S. A vector space model for automatic indexing, Communications of the ACM, 18, 11 (Nov. 1975). ACM Press, New York, NY, 1975, 613--620. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Van Rijsbergen, C. J., Information Retrieval, 2nd Ed. Butterworth-Heinemann, Newton, MA, 1979. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. The Gutenberg Project. Retrieved March 5, 2006, from http://www.gutenberg.orgGoogle ScholarGoogle Scholar

Index Terms

  1. An infrastructure for the evaluation and comparison of information retrieval systems

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      SIGITE '06: Proceedings of the 7th conference on Information technology education
      October 2006
      180 pages
      ISBN:1595935215
      DOI:10.1145/1168812

      Copyright © 2006 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 19 October 2006

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • Article

      Acceptance Rates

      Overall Acceptance Rate176of429submissions,41%
    • Article Metrics

      • Downloads (Last 12 months)4
      • Downloads (Last 6 weeks)0

      Other Metrics

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader