Sorry, I don't understand your search. ×
Back to Search Start Over

Boosting Scene Graph Generation with Visual Relation Saliency

Authors :
Yong Zhang
Yingwei Pan
Ting Yao
Rui Huang
Tao Mei
Chang-Wen Chen
Source :
ACM Transactions on Multimedia Computing, Communications, and Applications. 19:1-17
Publication Year :
2023
Publisher :
Association for Computing Machinery (ACM), 2023.

Abstract

The scene graph is a symbolic data structure that comprehensively describes the objects and visual relations in a visual scene, while ignoring the inherent perceptual saliency of each visual relation (i.e., relation saliency). However, humans often quickly allocate attention to important/salient visual relations in a scene. To align with such human perception of a scene, we explicitly model the perceptual saliency of visual relation in scene graph by upgrading each graph edge (i.e., visual relation) with an attribute of relation saliency. We present a new design, named as Saliency-guided Message Passing (SMP), that boosts the generation of such scene graph structure with the guidance from the visual relation saliency. Technically, an object interaction encoder is first utilized to strengthen object relation representations by jointly exploiting the appearance, semantic, and spatial relations in between. A branch is further leveraged to estimate the relation saliency of each visual relation by ordinal regression. Next, conditioned on the object and relation features (coupled with the estimated relation saliency), our SMP enhances scene graph generation by performing message passing over the objects and the most salient relations. Extensive experiments on VG-KR and VG150 datasets demonstrate the superiority of SMP for the scene graph generation. Moreover, we empirically validate the compelling generalizability of the learned scene graphs via SMP on downstream tasks like cross-model retrieval and image captioning.

Details

ISSN :
15516865 and 15516857
Volume :
19
Database :
OpenAIRE
Journal :
ACM Transactions on Multimedia Computing, Communications, and Applications
Accession number :
edsair.doi...........d413942def8ea932514801e243898c6c