Information Retrieval Evaluation (Record no. 85071)

000 -LEADER
fixed length control field 03787nam a22004575i 4500
001 - CONTROL NUMBER
control field 978-3-031-02276-0
005 - DATE AND TIME OF LATEST TRANSACTION
control field 20240730163853.0
008 - FIXED-LENGTH DATA ELEMENTS--GENERAL INFORMATION
fixed length control field 220601s2011 sz | s |||| 0|eng d
020 ## - INTERNATIONAL STANDARD BOOK NUMBER
ISBN 9783031022760
-- 978-3-031-02276-0
082 04 - CLASSIFICATION NUMBER
Call Number 004.6
100 1# - AUTHOR NAME
Author Harman, Donna.
245 10 - TITLE STATEMENT
Title Information Retrieval Evaluation
250 ## - EDITION STATEMENT
Edition statement 1st ed. 2011.
300 ## - PHYSICAL DESCRIPTION
Number of Pages XI, 107 p.
490 1# - SERIES STATEMENT
Series statement Synthesis Lectures on Information Concepts, Retrieval, and Services,
505 0# - FORMATTED CONTENTS NOTE
Remark 2 Introduction and Early History -- "Batch" Evaluation Since 1992 -- Interactive Evaluation -- Conclusion.
520 ## - SUMMARY, ETC.
Summary, etc Evaluation has always played a major role in information retrieval, with the early pioneers such as Cyril Cleverdon and Gerard Salton laying the foundations for most of the evaluation methodologies in use today. The retrieval community has been extremely fortunate to have such a well-grounded evaluation paradigm during a period when most of the human language technologies were just developing. This lecture has the goal of explaining where these evaluation methodologies came from and how they have continued to adapt to the vastly changed environment in the search engine world today. The lecture starts with a discussion of the early evaluation of information retrieval systems, starting with the Cranfield testing in the early 1960s, continuing with the Lancaster "user" study for MEDLARS, and presenting the various test collection investigations by the SMART project and by groups in Britain. The emphasis in this chapter is on the how and the why of the various methodologies developed. Thesecond chapter covers the more recent "batch" evaluations, examining the methodologies used in the various open evaluation campaigns such as TREC, NTCIR (emphasis on Asian languages), CLEF (emphasis on European languages), INEX (emphasis on semi-structured data), etc. Here again the focus is on the how and why, and in particular on the evolving of the older evaluation methodologies to handle new information access techniques. This includes how the test collection techniques were modified and how the metrics were changed to better reflect operational environments. The final chapters look at evaluation issues in user studies -- the interactive part of information retrieval, including a look at the search log studies mainly done by the commercial search engines. Here the goal is to show, via case studies, how the high-level issues of experimental design affect the final evaluations. Table of Contents: Introduction and Early History / "Batch" Evaluation Since 1992 / Interactive Evaluation/ Conclusion.
856 40 - ELECTRONIC LOCATION AND ACCESS
Uniform Resource Identifier https://doi.org/10.1007/978-3-031-02276-0
942 ## - ADDED ENTRY ELEMENTS (KOHA)
Koha item type eBooks
264 #1 -
-- Cham :
-- Springer International Publishing :
-- Imprint: Springer,
-- 2011.
336 ## -
-- text
-- txt
-- rdacontent
337 ## -
-- computer
-- c
-- rdamedia
338 ## -
-- online resource
-- cr
-- rdacarrier
347 ## -
-- text file
-- PDF
-- rda
650 #0 - SUBJECT ADDED ENTRY--SUBJECT 1
-- Computer networks .
650 14 - SUBJECT ADDED ENTRY--SUBJECT 1
-- Computer Communication Networks.
830 #0 - SERIES ADDED ENTRY--UNIFORM TITLE
-- 1947-9468
912 ## -
-- ZDB-2-SXSC

No items available.