论文标题
对眼科诊断的可解释深度学习方法的定量和定性评估
Quantitative and Qualitative Evaluation of Explainable Deep Learning Methods for Ophthalmic Diagnosis
论文作者
论文摘要
背景:缺乏对算法(例如深度学习)做出的决策的解释,尽管对多个问题有高度准确的结果,但仍阻碍了临床社区的接受。最近,出现了用于解释深度学习模型的归因方法,并且已经在医学成像问题上进行了测试。在标准的机器学习数据集而不是医学图像上比较归因方法的性能。在这项研究中,我们进行了比较分析,以确定视网膜OCT诊断的最合适的解释性方法。 方法:一种常用的深度学习模型(称为Inception V3)经过训练,可以诊断3种视网膜疾病 - 脉络膜新生血管形成(CNV)(CNV),糖尿病性黄斑水肿(DME)和DRUSEN。 14位临床医生的小组对13种不同归因方法的解释进行了临床意义。从临床医生那里获得了有关此类方法的当前和未来范围的反馈。 结果:一种基于泰勒级数扩展的归因方法,称为Deep Taylor,被中位评级为3.85/5的临床医生评为最高。随后是另外两种归因方法,指导的反向传播和摇摆(Shapley添加性解释)。 结论:深度学习模型的解释可以使它们在临床诊断中更透明。这项研究比较了视网膜OCT诊断背景下的不同解释方法,发现最佳性能方法可能不是认为其他深度学习任务的方法。总体而言,在研究中调查的临床医生受到了高度的接受。 关键字:可解释的AI,深度学习,机器学习,图像处理,光学连贯性层析成像,视网膜,糖尿病黄斑水肿,脉络膜新生血管化,drusen
Background: The lack of explanations for the decisions made by algorithms such as deep learning has hampered their acceptance by the clinical community despite highly accurate results on multiple problems. Recently, attribution methods have emerged for explaining deep learning models, and they have been tested on medical imaging problems. The performance of attribution methods is compared on standard machine learning datasets and not on medical images. In this study, we perform a comparative analysis to determine the most suitable explainability method for retinal OCT diagnosis. Methods: A commonly used deep learning model known as Inception v3 was trained to diagnose 3 retinal diseases - choroidal neovascularization (CNV), diabetic macular edema (DME), and drusen. The explanations from 13 different attribution methods were rated by a panel of 14 clinicians for clinical significance. Feedback was obtained from the clinicians regarding the current and future scope of such methods. Results: An attribution method based on a Taylor series expansion, called Deep Taylor was rated the highest by clinicians with a median rating of 3.85/5. It was followed by two other attribution methods, Guided backpropagation and SHAP (SHapley Additive exPlanations). Conclusion: Explanations of deep learning models can make them more transparent for clinical diagnosis. This study compared different explanations methods in the context of retinal OCT diagnosis and found that the best performing method may not be the one considered best for other deep learning tasks. Overall, there was a high degree of acceptance from the clinicians surveyed in the study. Keywords: explainable AI, deep learning, machine learning, image processing, Optical coherence tomography, retina, Diabetic macular edema, Choroidal Neovascularization, Drusen