Search results for key=MMM2001a :
1 match found.
Search my entire BibTeX database
Abstract |
BibTeX entry |
Postscript |
PDF |
Powerpoint |
Refereed full papers (journals, book chapters, international conferences)
2001
-
Henning Müller, Wolfgang Müller, Stéphane
Marchand-Maillet, David McG. Squire and Thierry Pun,
A web-based evaluation system for content-based image
retrieval,
In Proceedings of the 3rd International Workshop on
Multimedia Information Retrieval (in conjunction with ACM Multimedia
2001),
Ottawa, Canada, pp. 50-54, October 5 2001.
This papers describes a benchmark test for content-based
image retrieval systems (CBIRSs) with the query by example (QBE) query
paradigm. This benchmark is accessible via the Internet and thus allows
to evaluate any image retrieval system which is compliant with the
Multimedia Markup Language (MRML) for query formulation and result
transmission. Thus it allows a quick and easy comparison between
different features and algorithms for CBIRSs. The benchmark is not
only based on a standardized communication protocol to do the
communication between the benchmark server and the benchmarked system,
but it also uses a freely downloadable image database for the
evaluation to make the results reproducible. A CBIR system that uses
MRML and other components to develop MRML-based applications can be
downloaded free of charge as well. The evaluation is based on several
queries and known relevance sets for these queries. Several answer sets
for the same query image are possible if user judgments of several
users exist, thus almost any sort of user judgment can be incorporated
into the system. The final results are averaged over all the queries.
The evaluation of several steps of relevance feedback based on the
collected relevance judgments is also included into the benchmark. The
performance of relevance feedback is often regarded to be even more
important than the performance in the first query step because only
with relevance feedback the adaptation of the system to the users
subjective goal can be measured. For the evaluation of a system with
relevance feedback, the same evaluation measures are used on the query
results as for the first query step.
|