Partial-duplicate image retrieval using spatial and visual contextual clues Online publication date: Thu, 19-Mar-2020
by Wendi Sun; Tao Wang; Zhili Zhou
International Journal of Embedded Systems (IJES), Vol. 12, No. 2, 2020
Abstract: The traditional bag-of-visual-words (BOW) model quantifies the local features to the visual words to achieve efficient content-based image retrieval. However, since it causes considerable quantisation error and ignores the spatial relationships between visual words, the accuracy of partial-duplicate image retrieval based on BOW model is limited. In order to reduce the quantisation error and improve the discriminability of visual words, many partial-duplicate image retrieval methods have been proposed, which make use of the advantages of the geometric clues between visual words. In this paper, we propose a novel partial-duplicate scheme by using both spatial and visual contextual clues for removing the false matches effectively, which not only encodes the relationships of orientation, distance and dominant orientation between the referential visual word and its context, but also takes the colour information between visual words into consideration. Experimental results reveal that our proposed algorithm achieves superior performance to the state-of-the-art methods for partial-duplicate image retrieval.
Online publication date: Thu, 19-Mar-2020
If you are not a subscriber and you just want to read the full contents of this article, buy online access here.Complimentary Subscribers, Editors or Members of the Editorial Board of the International Journal of Embedded Systems (IJES):
Login with your Inderscience username and password:
Want to subscribe?
A subscription gives you complete access to all articles in the current issue, as well as to all articles in the previous three years (where applicable). See our Orders page to subscribe.
If you still need assistance, please email firstname.lastname@example.org