Narrow your search

Library

AP (2)

EhB (2)

KDG (2)

VUB (1)


Resource type

book (2)

digital (2)


Language

English (4)


Year
From To Submit

2010 (2)

2009 (2)

Listing 1 - 4 of 4
Sort by

Digital
ImageCLEF : Experimental Evaluation in Visual Information Retrieval
Authors: --- --- ---
ISBN: 9783642151811 9783642151804 9783642264634 9783642151828 Year: 2010 Publisher: Berlin, Heidelberg Springer

Loading...
Export citation

Choose an application

Bookmark

Abstract

The creation and consumption of content, especially visual content, is ingrained into our modern world. This book contains a collection of texts centered on the evaluation of image retrieval systems. To enable reproducible evaluation we must create standardized benchmarks and evaluation methodologies. The individual chapters in this book highlight major issues and challenges in evaluating image retrieval systems and describe various initiatives that provide researchers with the necessary evaluation resources. In particular they describe activities within ImageCLEF, an initiative to evaluate cross-language image retrieval systems which has been running as part of the Cross Language Evaluation Forum (CLEF) since 2003. To this end, the editors collected contributions from a range of people: those involved directly with ImageCLEF, such as the organizers of specific image retrieval or annotation tasks; participants who have developed techniques to tackle the challenges set forth by the organizers; and people from industry and academia involved with image retrieval and evaluation generally. Mostly written for researchers in academia and industry, the book stresses the importance of combing textual and visual information – a multimodal approach – for effective retrieval. It provides the reader with clear ideas about information retrieval and its evaluation in contexts and domains such as healthcare, robot vision, press photography, and the Web.


Book
ImageCLEF : Experimental Evaluation in Visual Information Retrieval
Authors: --- --- --- ---
ISBN: 9783642151811 9783642151804 9783642264634 9783642151828 Year: 2010 Publisher: Berlin Heidelberg Springer Berlin Heidelberg

Loading...
Export citation

Choose an application

Bookmark

Abstract

The creation and consumption of content, especially visual content, is ingrained into our modern world. This book contains a collection of texts centered on the evaluation of image retrieval systems. To enable reproducible evaluation we must create standardized benchmarks and evaluation methodologies. The individual chapters in this book highlight major issues and challenges in evaluating image retrieval systems and describe various initiatives that provide researchers with the necessary evaluation resources. In particular they describe activities within ImageCLEF, an initiative to evaluate cross-language image retrieval systems which has been running as part of the Cross Language Evaluation Forum (CLEF) since 2003. To this end, the editors collected contributions from a range of people: those involved directly with ImageCLEF, such as the organizers of specific image retrieval or annotation tasks; participants who have developed techniques to tackle the challenges set forth by the organizers; and people from industry and academia involved with image retrieval and evaluation generally. Mostly written for researchers in academia and industry, the book stresses the importance of combing textual and visual information - a multimodal approach - for effective retrieval. It provides the reader with clear ideas about information retrieval and its evaluation in contexts and domains such as healthcare, robot vision, press photography, and the Web.


Digital
Evaluating Systems for Multilingual and Multimodal Information Access : 9th Workshop of the Cross-Language Evaluation Forum, CLEF 2008, Aarhus, Denmark, September 17-19, 2008, Revised Selected Papers
Authors: --- --- --- --- --- et al.
ISBN: 9783642044472 Year: 2009 Publisher: Berlin, Heidelberg Springer Berlin Heidelberg


Book
Evaluating Systems for Multilingual and Multimodal Information Access : 9th Workshop of the Cross-Language Evaluation Forum, CLEF 2008, Aarhus, Denmark, September 17-19, 2008, Revised Selected Papers
Authors: --- --- --- --- --- et al.
ISBN: 9783642044472 Year: 2009 Publisher: Berlin Heidelberg Springer Berlin Heidelberg

Loading...
Export citation

Choose an application

Bookmark

Abstract

The ninth campaign of the Cross-Language Evaluation Forum (CLEF) for European languages was held from January to September 2008. There were seven main eval- tion tracks in CLEF 2008 plus two pilot tasks. The aim, as usual, was to test the p- formance of a wide range of multilingual information access (MLIA) systems or s- tem components. This year, 100 groups, mainly but not only from academia, parti- pated in the campaign. Most of the groups were from Europe but there was also a good contingent from North America and Asia plus a few participants from South America and Africa. Full details regarding the design of the tracks, the methodologies used for evaluation, and the results obtained by the participants can be found in the different sections of these proceedings. The results of the CLEF 2008 campaign were presented at a two-and-a-half day workshop held in Aarhus, Denmark, September 17-19, and attended by 150 resear- ers and system developers. The annual workshop, held in conjunction with the European Conference on Digital Libraries, plays an important role by providing the opportunity for all the groups that have participated in the evaluation campaign to get together comparing approaches and exchanging ideas. The schedule of the workshop was divided between plenary track overviews, and parallel, poster and breakout sessions presenting this year's experiments and discu- ing ideas for the future. There were several invited talks.

Listing 1 - 4 of 4
Sort by