论文标题

可解释的多模式融合网络揭示了脑认知的机制

Interpretable multimodal fusion networks reveal mechanisms of brain cognition

论文作者

Hu, Wenxing, Meng, Xianghe, Bai, Yuntong, Zhang, Aiying, Cai, Biao, Zhang, Gemeng, Wilson, Tony W., Stephen, Julia M., Calhoun, Vince D., Wang, Yu-Ping

论文摘要

多模式融合通过提供更全面的观点来使疾病诊断受益。由于数据异质性以及模式内和模式之间的复杂性,开发算法具有挑战性。已经开发了基于深网的数据融合模型来捕获复杂的关联,并且诊断的性能得到了相应的提高。超越诊断预测,对疾病机制的评估对于生物医学研究至关重要。但是,基于深网的数据融合模型很难解释,这给研究生物学机制带来了困难。在这项工作中,我们开发了一种可解释的多模式融合模型,即GCAM-CCL,该模型可以同时执行自动诊断和结果解释。 GCAM-CCL模型可以生成可解释的激活图,该图量化了输入特征的像素级贡献。这是通过使用基于梯度的权重结合中间特征图来实现的。此外,估计的激活图是特定于类的激活图,并且捕获的跨数据关联与兴趣/标签相关,这进一步促进了特定于类的分析和生物学机制分析。我们在脑成像基因研究中验证了GCAM-CCL模型,并显示GCAM-CCL在分类和机理分析方面表现良好。机理分析表明,在任务-FMRI扫描中,首先激活了几个与对象识别相关的兴趣区域(ROI),然后参与了几个下游编码ROI。结果还表明,较高的认知性能组可能具有更强的神经传递信号传导,而较低的认知性能组可能在脑/神经元发育中存在问题,这是由于遗传变异所致。

Multimodal fusion benefits disease diagnosis by providing a more comprehensive perspective. Developing algorithms is challenging due to data heterogeneity and the complex within- and between-modality associations. Deep-network-based data-fusion models have been developed to capture the complex associations and the performance in diagnosis has been improved accordingly. Moving beyond diagnosis prediction, evaluation of disease mechanisms is critically important for biomedical research. Deep-network-based data-fusion models, however, are difficult to interpret, bringing about difficulties for studying biological mechanisms. In this work, we develop an interpretable multimodal fusion model, namely gCAM-CCL, which can perform automated diagnosis and result interpretation simultaneously. The gCAM-CCL model can generate interpretable activation maps, which quantify pixel-level contributions of the input features. This is achieved by combining intermediate feature maps using gradient-based weights. Moreover, the estimated activation maps are class-specific, and the captured cross-data associations are interest/label related, which further facilitates class-specific analysis and biological mechanism analysis. We validate the gCAM-CCL model on a brain imaging-genetic study, and show gCAM-CCL's performed well for both classification and mechanism analysis. Mechanism analysis suggests that during task-fMRI scans, several object recognition related regions of interests (ROIs) are first activated and then several downstream encoding ROIs get involved. Results also suggest that the higher cognition performing group may have stronger neurotransmission signaling while the lower cognition performing group may have problem in brain/neuron development, resulting from genetic variations.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源