Back to Search
Start Over
Saliency-Aware Nonparametric Foreground Annotation Based on Weakly Labeled Data.
- Source :
-
IEEE Transactions on Neural Networks & Learning Systems . Jun2016, Vol. 27 Issue 6, p1253-1265. 13p. - Publication Year :
- 2016
-
Abstract
- In this paper, we focus on annotating the foreground of an image. More precisely, we predict both image-level labels (category labels) and object-level labels (locations) for objects within a target image in a unified framework. Traditional learning-based image annotation approaches are cumbersome, because they need to establish complex mathematical models and be frequently updated as the scale of training data varies considerably. Thus, we advocate the nonparametric method, which has shown potential in numerous applications and turned out to be attractive thanks to its advantages, i.e., lightweight training load and scalability. In particular, we exploit the salient object windows to describe images, which is beneficial to image retrieval and, thus, the subsequent image-level annotation and localization tasks. Our method, namely, saliency-aware nonparametric foreground annotation, is practical to alleviate the full label requirement of training data, and effectively addresses the problem of foreground annotation. The proposed method only relies on retrieval results from the image database, while pretrained object detectors are no longer necessary. Experimental results on the challenging PASCAL VOC 2007 and PASCAL VOC 2008 demonstrate the advance of our method. [ABSTRACT FROM PUBLISHER]
Details
- Language :
- English
- ISSN :
- 2162237X
- Volume :
- 27
- Issue :
- 6
- Database :
- Academic Search Index
- Journal :
- IEEE Transactions on Neural Networks & Learning Systems
- Publication Type :
- Periodical
- Accession number :
- 115559438
- Full Text :
- https://doi.org/10.1109/TNNLS.2015.2488637