Back to Search Start Over

W-tree indexing for fast visual word generation.

Authors :
Shi M
Xu R
Tao D
Xu C
Source :
IEEE transactions on image processing : a publication of the IEEE Signal Processing Society [IEEE Trans Image Process] 2013 Mar; Vol. 22 (3), pp. 1209-22. Date of Electronic Publication: 2012 Nov 20.
Publication Year :
2013

Abstract

The bag-of-visual-words representation has been widely used in image retrieval and visual recognition. The most time-consuming step in obtaining this representation is the visual word generation, i.e., assigning visual words to the corresponding local features in a high-dimensional space. Recently, structures based on multibranch trees and forests have been adopted to reduce the time cost. However, these approaches cannot perform well without a large number of backtrackings. In this paper, by considering the spatial correlation of local features, we can significantly speed up the time consuming visual word generation process while maintaining accuracy. In particular, visual words associated with certain structures frequently co-occur; hence, we can build a co-occurrence table for each visual word for a large-scale data set. By associating each visual word with a probability according to the corresponding co-occurrence table, we can assign a probabilistic weight to each node of a certain index structure (e.g., a KD-tree and a K-means tree), in order to re-direct the searching path to be close to its global optimum within a small number of backtrackings. We carefully study the proposed scheme by comparing it with the fast library for approximate nearest neighbors and the random KD-trees on the Oxford data set. Thorough experimental results suggest the efficiency and effectiveness of the new scheme.

Details

Language :
English
ISSN :
1941-0042
Volume :
22
Issue :
3
Database :
MEDLINE
Journal :
IEEE transactions on image processing : a publication of the IEEE Signal Processing Society
Publication Type :
Academic Journal
Accession number :
23192558
Full Text :
https://doi.org/10.1109/TIP.2012.2228494