Back to Search Start Over

EmbraceNet: A robust deep learning architecture for multimodal classification

Authors :
Choi, Jun-Ho
Lee, Jong-Seok
Source :
Information Fusion 51 (2019) 259-270
Publication Year :
2019

Abstract

Classification using multimodal data arises in many machine learning applications. It is crucial not only to model cross-modal relationship effectively but also to ensure robustness against loss of part of data or modalities. In this paper, we propose a novel deep learning-based multimodal fusion architecture for classification tasks, which guarantees compatibility with any kind of learning models, deals with cross-modal information carefully, and prevents performance degradation due to partial absence of data. We employ two datasets for multimodal classification tasks, build models based on our architecture and other state-of-the-art models, and analyze their performance on various situations. The results show that our architecture outperforms the other multimodal fusion architectures when some parts of data are not available.<br />Comment: Code available at https://github.com/idearibosome/embracenet

Details

Database :
arXiv
Journal :
Information Fusion 51 (2019) 259-270
Publication Type :
Report
Accession number :
edsarx.1904.09078
Document Type :
Working Paper
Full Text :
https://doi.org/10.1016/j.inffus.2019.02.010