Multiple representations, similarity matching, and results fusion for content-based image retrieval |
| |
Authors: | Noureddine Abbadeni |
| |
Affiliation: | (1) Faculté des sciences, Département d'informatique, Université de Sherbrooke, Sherbrooke, QC, J1K 2R1, Canada |
| |
Abstract: | In this paper, we show how the use of multiple content representations and their fusion can improve the performance of content-based
image retrieval systems. We consider the case of texture and propose a new algorithm for texture retrieval based on multiple
representations and their results fusion. Texture content is modeled using two different models: the well-known autoregressive
model and a perceptual model based on perceptual features such as coarseness and directionality. In the case of the perceptual
model, two viewpoints are considered: perceptual features are computed based on the original images viewpoint and on the autocovariance
function viewpoint (corresponding to original images). So we consider a total of three content representations. The similarity
measure used is based on Gower's index of similarity. Simple results of the fusion models are used to merge search results
returned by different representations. Experimentations and benchmarking carried out on the well-known Brodatz database show
a drastic improvement in search effectiveness with the fused model without necessarily altering their efficiency in an important
way. |
| |
Keywords: | Content-based image retrieval Multiple representations Perceptual model Autoregressive model Similarity matching Results fusion |
本文献已被 SpringerLink 等数据库收录! |
|