University of Wollongong
Browse

Perceived similarity and visual descriptions in content-based image retrieval

Download (889.07 kB)
conference contribution
posted on 2024-11-14, 10:54 authored by Yuan Zhong, Lei Ye, Wanqing LiWanqing Li, Philip OgunbonaPhilip Ogunbona
The use of low-level feature descriptors is pervasive in content-based image retrieval tasks and the answer to the question of how well these features describe users’ intention is inconclusive. In this paper we devise experiments to gauge the degree of alignment between the description of target images by humans and that implicitly provided by low-level image feature descriptors. Data was collected on how humans perceive similarity in images. Using images judged by humans to be similar, as ground truth, the performance of some MPEG-7 visual feature descriptors were evaluated. It is found that various descriptors play different roles in different queries and their appropriate combination can improve the performance of retrieval tasks. This forms a basis for the development of adaptive weight assignment to features depending on the query and retrieval task.

History

Citation

Zhong, Y., Ye, L., Li, W. & Ogunbona, P. (2007). Perceived similarity and visual descriptions in content-based image retrieval. The IEEE International Symposium on Multimedia (pp. 173-180). IEEE Computer Society Press.

Parent title

ISM WORKSHOPS 2007: NINTH IEEE INTERNATIONAL SYMPOSIUM ON MULTIMEDIA - WORKSHOPS, PROCEEDINGS

Pagination

173-180

Language

English

RIS ID

25160

Usage metrics

    Categories

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC