Back to Search Start Over

Saliency-Aware Nonparametric Foreground Annotation Based on Weakly Labeled Data.

Authors :
Cao, Xiaochun
Zhang, Changqing
Fu, Huazhu
Guo, Xiaojie
Tian, Qi
Source :
IEEE Transactions on Neural Networks & Learning Systems. Jun2016, Vol. 27 Issue 6, p1253-1265. 13p.
Publication Year :
2016

Abstract

In this paper, we focus on annotating the foreground of an image. More precisely, we predict both image-level labels (category labels) and object-level labels (locations) for objects within a target image in a unified framework. Traditional learning-based image annotation approaches are cumbersome, because they need to establish complex mathematical models and be frequently updated as the scale of training data varies considerably. Thus, we advocate the nonparametric method, which has shown potential in numerous applications and turned out to be attractive thanks to its advantages, i.e., lightweight training load and scalability. In particular, we exploit the salient object windows to describe images, which is beneficial to image retrieval and, thus, the subsequent image-level annotation and localization tasks. Our method, namely, saliency-aware nonparametric foreground annotation, is practical to alleviate the full label requirement of training data, and effectively addresses the problem of foreground annotation. The proposed method only relies on retrieval results from the image database, while pretrained object detectors are no longer necessary. Experimental results on the challenging PASCAL VOC 2007 and PASCAL VOC 2008 demonstrate the advance of our method. [ABSTRACT FROM PUBLISHER]

Details

Language :
English
ISSN :
2162237X
Volume :
27
Issue :
6
Database :
Academic Search Index
Journal :
IEEE Transactions on Neural Networks & Learning Systems
Publication Type :
Periodical
Accession number :
115559438
Full Text :
https://doi.org/10.1109/TNNLS.2015.2488637