Back to Search Start Over

Learning a Combined Model of Visual Saliency for Fixation Prediction

Authors :
Jingwei Wang
C.-C. Jay Kuo
Ali Borji
Laurent Itti
Source :
IEEE Transactions on Image Processing. 25:1566-1579
Publication Year :
2016
Publisher :
Institute of Electrical and Electronics Engineers (IEEE), 2016.

Abstract

A large number of saliency models, each based on a different hypothesis, have been proposed over the past 20 years. In practice, while subscribing to one hypothesis or computational principle makes a model that performs well on some types of images, it hinders the general performance of a model on arbitrary images and large-scale data sets. One natural approach to improve overall saliency detection accuracy would then be fusing different types of models. In this paper, inspired by the success of late-fusion strategies in semantic analysis and multi-modal biometrics, we propose to fuse the state-of-the-art saliency models at the score level in a para-boosting learning fashion. First, saliency maps generated by several models are used as confidence scores. Then, these scores are fed into our para-boosting learner (i.e., support vector machine, adaptive boosting, or probability density estimator) to generate the final saliency map. In order to explore the strength of para-boosting learners, traditional transformation-based fusion strategies, such as Sum , Min , and Max , are also explored and compared in this paper. To further reduce the computation cost of fusing too many models, only a few of them are considered in the next step. Experimental results show that score-level fusion outperforms each individual model and can further reduce the performance gap between the current models and the human inter-observer model.

Details

ISSN :
19410042 and 10577149
Volume :
25
Database :
OpenAIRE
Journal :
IEEE Transactions on Image Processing
Accession number :
edsair.doi.dedup.....b682a80f0ae225ab914c7f7f2126f34c
Full Text :
https://doi.org/10.1109/tip.2016.2522380