Abstract
The paper introduces a model for interactive image retrieval utilising the geometrical framework of information retrieval (IR). We tackle the problem of image retrieval based on an expressive user information need in form of a textual-visual query, where a user is attempting to find an image similar to the picture in their mind during querying. The user information need is expressed using guided visual feedback based on Information Foraging which lets the user perception embed within the model via semantic Hilbert space (SHS). This framework is based on the mathematical formalism of quantum probabilities and aims to understand the relationship between user textual and image input, where the image in the input is considered a form of visual feedback. We propose SHS, a quantum-inspired approach where the textual-visual query is regarded analogously to a physical system that allows for modelling different system states and their dynamic changes thereof based on observations (such as queries, relevance judgements). We will be able to learn the input multimodal representation and relationships between textual-image queries for retrieving images. Our experiments are conducted on the MIT States and Fashion200k datasets that demonstrate the effectiveness of finding particular images autocratically when the user inputs are semantically expressive.Citation
Jaiswal AK., Liu H, Frommholz I (2021) 'Semantic Hilbert space for interactive image retrieval', 2021 ACM SIGIR International Conference on Theory of Information Retrieval - Online, Association for Computing Machinery, Inc.Publisher
Association for Computing Machinery, IncAdditional Links
https://dl.acm.org/doi/10.1145/3471158.3472253Type
Conference papers, meetings and proceedingsLanguage
enISBN
9781450386111ae974a485f413a2113503eed53cd6c53
10.1145/3471158.3472253