1. Imaging-genomic spatial-modality attentive fusion for studying neuropsychiatric disorders.
- Author
-
Rahaman, Md, Garg, Yash, Iraji, Armin, Fu, Zening, Kochunov, Peter, Hong, L, Van Erp, Theo, Preda, Adrian, Chen, Jiayu, and Calhoun, Vince
- Subjects
bottleneck attention module (BAM) ,functional network connectivity (FNC) ,imaging‐genetics ,multimodal fusion ,resting‐state functional and structural MRI ,schizophrenia classification ,single‐nucleotide polymorphism (SNP) ,Humans ,Schizophrenia ,Magnetic Resonance Imaging ,Deep Learning ,Adult ,Multimodal Imaging ,Neuroimaging ,Attention ,Brain ,Neural Networks ,Computer ,Machine Learning - Abstract
Multimodal learning has emerged as a powerful technique that leverages diverse data sources to enhance learning and decision-making processes. Adapting this approach to analyzing data collected from different biological domains is intuitive, especially for studying neuropsychiatric disorders. A complex neuropsychiatric disorder like schizophrenia (SZ) can affect multiple aspects of the brain and biologies. These biological sources each present distinct yet correlated expressions of subjects underlying physiological processes. Joint learning from these data sources can improve our understanding of the disorder. However, combining these biological sources is challenging for several reasons: (i) observations are domain specific, leading to data being represented in dissimilar subspaces, and (ii) fused data are often noisy and high-dimensional, making it challenging to identify relevant information. To address these challenges, we propose a multimodal artificial intelligence model with a novel fusion module inspired by a bottleneck attention module. We use deep neural networks to learn latent space representations of the input streams. Next, we introduce a two-dimensional (spatio-modality) attention module to regulate the intermediate fusion for SZ classification. We implement spatial attention via a dilated convolutional neural network that creates large receptive fields for extracting significant contextual patterns. The resulting joint learning framework maximizes complementarity allowing us to explore the correspondence among the modalities. We test our model on a multimodal imaging-genetic dataset and achieve an SZ prediction accuracy of 94.10% (p
- Published
- 2024