skip to main content
10.1145/3020165.3022154acmconferencesArticle/Chapter ViewAbstractPublication PagesirConference Proceedingsconference-collections
short-paper

Identifying Useful and Important Information within Retrieved Documents

Published:07 March 2017Publication History

ABSTRACT

We describe an initial study into the identification of important and useful information units within documents retrieved by an information retrieval system in response to a user query created in response to an underlying information need. This study is part of a large investigation of the exploitation of useful and important units from retrieved documents to generate rich document surrogates to improve user search experience. We report three user studies using a crowdsourcing platform, where participants were first asked to read an information need and contents of a relevant document and then to perform actions depending on the type of study: i) write important information units (WIIU), ii) highlight important information units (HIIU) and iii) assess importance of already highlighted information units (AIHIU). Further, we discuss a novel mechanism of measuring similarities between content annotations. We find majority agreement of about 0.489 and pairwise agreement of 0.340 among users annotation in the AIHIU study, and average cosine similarity of 0.50 and 0.57 between participant annotations and documents in the WIIU and HIIU studies respectively.

References

  1. A. Al-Maskari and M. Sanderson. A review of factors influencing user satisfaction in information retrieval. JASIST, 2010, 61(5):859--868. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. P. Arora and G. J. F. Jones. Position paper: Promoting user engagement and learning in search tasks by effective document representation. In Proceedings of SAL workshop, SIGIR 2016 .Google ScholarGoogle Scholar
  3. M. Cole, J. Liu, N. Belkin, R. Bierig, J. Gwizdka, C. Liu, J. Zhang, and X. Zhang. Usefulness as the criterion for evaluation of interactive information retrieval. Proc. HCIR , pages 1--4, 2009.Google ScholarGoogle Scholar
  4. I. Habernal, M. Sukhareva, F. Raiber, A. Shtok, O. Kurland, H. Ronen, J. Bar-Ilan, and I. Gurevych. New collection announcement: Focused retrieval over the web. In Proceedings of SIGIR 2016, pages 701--704. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. K. Järvelin and J. Kekäläinen. Ir evaluation methods for retrieving highly relevant documents. In Proceedings of SIGIR 2000, pages 41--48. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. J. Kamps, J. Pehcevski, G. Kazai, M. Lalmas, and S. Robertson. Inex 2007 evaluation measures. In International Workshop of the Initiative for the Evaluation of XML Retrieval, pages 24--33. Springer, 2007.Google ScholarGoogle Scholar
  7. E. Kanoulas, B. Carterette, M. Hall, P. Clough, and M. Sanderson. Overview of the TREC 2012 Session Track. 2012.Google ScholarGoogle Scholar
  8. M. P. Kato, M. Ekstrand-Abueg, V. Pavlu, T. Sakai, T. Yamamoto, and M. Iwata. Overview of the ntcir-11 mobileclick task. In NTCIR, 2014.Google ScholarGoogle Scholar
  9. D. Kelly and C. Cool. The effects of topic familiarity on information search behavior. JCDL '02, pages 74--75. ACM, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. J. Mao, Y. Liu, K. Zhou, J.-Y. Nie, J. Song, M. Zhang, S. Ma, J. Sun, and H. Luo. When does relevance mean usefulness and user satisfaction in web search? In Proceedings of SIGIR 2016, pages 463--472. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. T. Mikolov, K. Chen, G. Corrado, and J. Dean. Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781, 2013.Google ScholarGoogle Scholar
  12. T. Saracevic. Relevance: A review of the literature and a framework for thinking on the notion in information science. part iii: Behavior and effects of relevance. JASIST, 2007, 58(13):2126--2144. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. M. Shokouhi and Q. Guo. From queries to cards: Re-ranking proactive card recommendations based on reactive search history. In Proceedings of SIGIR 2015, pages 695--704. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. A. Trotman and S. Geva. Passage retrieval and other xml-retrieval tasks. In Proceedings of the SIGIR 2006 Workshop on XML Element Retrieval Methodology, pages 43--50, 2006.Google ScholarGoogle Scholar
  15. R. W. White, J. M. Jose, and I. Ruthven. Using top-ranking sentences to facilitate effective information access. JASIST, 2005, 56(10):1113--1125. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. E. Yilmaz, M. Verma, N. Craswell, F. Radlinski, and P. Bailey. Relevance and effort: An analysis of document utility. In Proceedings of CIKM 2014, pages 91--100. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Identifying Useful and Important Information within Retrieved Documents

              Recommendations

              Comments

              Login options

              Check if you have access through your login credentials or your institution to get full access on this article.

              Sign in
              • Published in

                cover image ACM Conferences
                CHIIR '17: Proceedings of the 2017 Conference on Conference Human Information Interaction and Retrieval
                March 2017
                454 pages
                ISBN:9781450346771
                DOI:10.1145/3020165
                • Conference Chairs:
                • Ragnar Nordlie,
                • Nils Pharo,
                • Program Chairs:
                • Luanne Freund,
                • Birger Larsen,
                • Dan Russel

                Copyright © 2017 ACM

                Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

                Publisher

                Association for Computing Machinery

                New York, NY, United States

                Publication History

                • Published: 7 March 2017

                Permissions

                Request permissions about this article.

                Request Permissions

                Check for updates

                Qualifiers

                • short-paper

                Acceptance Rates

                CHIIR '17 Paper Acceptance Rate10of48submissions,21%Overall Acceptance Rate55of163submissions,34%

              PDF Format

              View or Download as a PDF file.

              PDF

              eReader

              View online with eReader.

              eReader