Search results for key=MMM2001a : 1 match found.

Search my entire BibTeX database
Output format: Text
BibTeX entry
     Combine using:
AND OR

Abstract icon Abstract BibTeX icon BibTeX entry Postscript icon Postscript PDF icon PDF PPT icon Powerpoint

Refereed full papers (journals, book chapters, international conferences)

2001

  • Henning Müller, Wolfgang Müller, Stéphane Marchand-Maillet, David McG. Squire and Thierry Pun, A web-based evaluation system for content-based image retrieval, In Proceedings of the 3rd International Workshop on Multimedia Information Retrieval (in conjunction with ACM Multimedia 2001), Ottawa, Canada, pp. 50-54, October 5 2001.

    This papers describes a benchmark test for content-based image retrieval systems (CBIRSs) with the query by example (QBE) query paradigm. This benchmark is accessible via the Internet and thus allows to evaluate any image retrieval system which is compliant with the Multimedia Markup Language (MRML) for query formulation and result transmission. Thus it allows a quick and easy comparison between different features and algorithms for CBIRSs. The benchmark is not only based on a standardized communication protocol to do the communication between the benchmark server and the benchmarked system, but it also uses a freely downloadable image database for the evaluation to make the results reproducible. A CBIR system that uses MRML and other components to develop MRML-based applications can be downloaded free of charge as well. The evaluation is based on several queries and known relevance sets for these queries. Several answer sets for the same query image are possible if user judgments of several users exist, thus almost any sort of user judgment can be incorporated into the system. The final results are averaged over all the queries. The evaluation of several steps of relevance feedback based on the collected relevance judgments is also included into the benchmark. The performance of relevance feedback is often regarded to be even more important than the performance in the first query step because only with relevance feedback the adaptation of the system to the users subjective goal can be measured. For the evaluation of a system with relevance feedback, the same evaluation measures are used on the query results as for the first query step.