Back to Search Start Over

Interpretable Visual Understanding with Cognitive Attention Network

Authors :
Tang, Xuejiao
Zhang, Wenbin
Yu, Yi
Turner, Kea
Derr, Tyler
Wang, Mengyu
Ntoutsi, Eirini
Publication Year :
2021

Abstract

While image understanding on recognition-level has achieved remarkable advancements, reliable visual scene understanding requires comprehensive image understanding on recognition-level but also cognition-level, which calls for exploiting the multi-source information as well as learning different levels of understanding and extensive commonsense knowledge. In this paper, we propose a novel Cognitive Attention Network (CAN) for visual commonsense reasoning to achieve interpretable visual understanding. Specifically, we first introduce an image-text fusion module to fuse information from images and text collectively. Second, a novel inference module is designed to encode commonsense among image, query and response. Extensive experiments on large-scale Visual Commonsense Reasoning (VCR) benchmark dataset demonstrate the effectiveness of our approach. The implementation is publicly available at https://github.com/tanjatang/CAN<br />Comment: ICANN21

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2108.02924
Document Type :
Working Paper