Contextual Image Annotation via Projection and Quantum Theory Inspired Measurement for Integration of Text and Visual Features
- Cite this paper as:
- Kaliciak L., Wang J., Song D., Zhang P., Hou Y. (2011) Contextual Image Annotation via Projection and Quantum Theory Inspired Measurement for Integration of Text and Visual Features. In: Song D., Melucci M., Frommholz I., Zhang P., Wang L., Arafat S. (eds) Quantum Interaction. QI 2011. Lecture Notes in Computer Science, vol 7052. Springer, Berlin, Heidelberg
Multimedia information retrieval suffers from the semantic gap, a difference between human perception and machine representation of images. In order to reduce the gap, a quantum theory inspired theoretical framework for integration of text and visual features has been proposed. This article is a follow-up work on this model. Previously, two relatively straightforward statistical approaches for making associations between dimensions of both feature spaces were employed, but with unsatisfactory results. In this paper, we propose to alleviate the problem regarding unannotated images by projecting them onto subspaces representing visual context and by incorporating a quantum-like measurement. The proposed principled approach extends the traditional vector space model (VSM) and seamlessly integrates with the tensor-based framework. Here, we experimentally test the novel association methods in a small-scale experiment.
Keywordsmultimedia retrieval quantum theory image annotation tensor product
Unable to display preview. Download preview PDF.