Back to Search Start Over

Hierarchical Cross-Modality Semantic Correlation Learning Model for Multimodal Summarization

Authors :
Zhang, Litian
Zhang, Xiaoming
Pan, Junshu
Huang, Feiran
Publication Year :
2021

Abstract

Multimodal summarization with multimodal output (MSMO) generates a summary with both textual and visual content. Multimodal news report contains heterogeneous contents, which makes MSMO nontrivial. Moreover, it is observed that different modalities of data in the news report correlate hierarchically. Traditional MSMO methods indistinguishably handle different modalities of data by learning a representation for the whole data, which is not directly adaptable to the heterogeneous contents and hierarchical correlation. In this paper, we propose a hierarchical cross-modality semantic correlation learning model (HCSCL) to learn the intra- and inter-modal correlation existing in the multimodal data. HCSCL adopts a graph network to encode the intra-modal correlation. Then, a hierarchical fusion framework is proposed to learn the hierarchical correlation between text and images. Furthermore, we construct a new dataset with relevant image annotation and image object label information to provide the supervision information for the learning procedure. Extensive experiments on the dataset show that HCSCL significantly outperforms the baseline methods in automatic summarization metrics and fine-grained diversity tests.<br />Comment: Accepted by AAAI2022

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2112.12072
Document Type :
Working Paper