可解释人工智能在脑肿瘤的应用概述
A Review of the Applications of Explainable Artificial Intelligence in Brain Tumors
DOI: 10.12677/acm.2026.163821, PDF,    科研立项经费支持
作者: 林国鉴, 刘书勇:绍兴文理学院医学院,浙江 绍兴;李珍珠, 刘志鹏, 油亚倩, 王波定*:宁波市第二医院神经外科,浙江 宁波
关键词: 脑肿瘤可解释人工智能(XAI)深度学习Brain Tumor Explainable Artificial Intelligence Deep Learning
摘要: 脑肿瘤作为一种侵袭性强、复发率高的疾病,对患者的生存率和生活质量有着显著影响。现代医学影像技术如MRI和CT为脑肿瘤的检测和诊断提供了宝贵的支持,而近年来兴起的人工智能(AI)和深度学习技术更是推动了脑肿瘤分析的自动化与智能化发展。然而,深度学习模型的“黑箱”性质限制了它们在实际医疗环境中的广泛应用,临床医生难以理解或信任这些缺乏解释性的模型。可解释人工智能(XAI)应运而生,旨在提升深度学习模型的透明性和可理解性,从而增强其在脑肿瘤检测、分类和预后预测中的应用效果。本文综述了当前XAI在脑肿瘤诊断中的应用进展,探讨了不同XAI方法的优缺点及其在临床场景中的适用性,总结了XAI在医学影像分析中的挑战与未来发展方向。
Abstract: Brain tumors are highly aggressive and have high recurrence rates, significantly affecting patients’ survival and quality of life. Modern medical imaging technologies, such as magnetic resonance imaging (MRI) and computed tomography (CT), provide valuable support for the detection and diagnosis of brain tumors. In recent years, the emergence of artificial intelligence (AI) and deep learning has further promoted the automation and intelligent analysis of brain tumors. However, the “black-box” nature of deep learning models limits their widespread adoption in real clinical environments, as clinicians find it difficult to understand or trust models that lack interpretability. Explainable artificial intelligence (XAI) has therefore emerged to enhance the transparency and interpretability of deep learning models, thereby improving their applicability in brain tumor detection, classification, and prognosis prediction. This article reviews the current progress of XAI in brain tumor diagnosis, discusses the advantages and limitations of different XAI methods and their suitability for clinical scenarios, and summarizes the challenges and future directions of XAI in medical imaging analysis.
文章引用:林国鉴, 李珍珠, 刘志鹏, 油亚倩, 刘书勇, 王波定. 可解释人工智能在脑肿瘤的应用概述[J]. 临床医学进展, 2026, 16(3): 546-553. https://doi.org/10.12677/acm.2026.163821

参考文献

[1] Aldape, K., Brindle, K.M., Chesler, L., Chopra, R., Gajjar, A., Gilbert, M.R., et al. (2019) Challenges to Curing Primary Brain Tumours. Nature Reviews Clinical Oncology, 16, 509-520. [Google Scholar] [CrossRef] [PubMed]
[2] Hu, L.S., Hawkins-Daarud, A., Wang, L., Li, J. and Swanson, K.R. (2020) Imaging of Intratumoral Heterogeneity in High-Grade Glioma. Cancer Letters, 477, 97-106. [Google Scholar] [CrossRef] [PubMed]
[3] Dorfner, F.J., Patel, J.B., Kalpathy-Cramer, J., Gerstner, E.R. and Bridge, C.P. (2025) A Review of Deep Learning for Brain Tumor Analysis in MRI. NPJ Precision Oncology, 9, Article No. 2. [Google Scholar] [CrossRef] [PubMed]
[4] Ottoni, M., Kasperczuk, A. and Tavora, L.M.N. (2025) Machine Learning in MRI Brain Imaging: A Review of Methods, Challenges, and Future Directions. Diagnostics, 15, Article No. 2692. [Google Scholar] [CrossRef
[5] Abraham, L.A., Palanisamy, G., Veerapu, G. and Nisha, J.S. (2025) Exploring the Potential of Explainable AI in Brain Tumor Detection and Classification: A Systematic Review. Artificial Intelligence Review, 59, Article No. 14. [Google Scholar] [CrossRef
[6] 陈园琼, 朱承璋. 医学影像处理的深度学习可解释性研究进展[J]. 浙江大学学报(理学版), 2021, 48(1): 18-29.
[7] Erukude, S.T., Chaitanya Marella, V. and Veluru, S.R. (2025) Explainable Deep Learning in Medical Imaging: Brain Tumor and Pneumonia Detection. 2025 4th International Conference on Innovative Mechanisms for Industry Applications (ICIMIA), Tirupur, 3-5 September 2025, 906-911. [Google Scholar] [CrossRef
[8] Bhati, D., Neha, F. and Amiruzzaman, M. (2024) A Survey on Explainable Artificial Intelligence (XAI) Techniques for Visualizing Deep Learning Models in Medical Imaging. Journal of Imaging, 10, Article No. 239. [Google Scholar] [CrossRef] [PubMed]
[9] Akgündoğdu, A. and Çelikbaş, Ş. (2025) Explainable Deep Learning Framework for Brain Tumor Detection: Integrating LIME, Grad-CAM, and SHAP for Enhanced Accuracy. Medical Engineering & Physics, 144, Article ID: 104405. [Google Scholar] [CrossRef
[10] Yan, F., Chen, Y., Xia, Y., Wang, Z. and Xiao, R. (2023) An Explainable Brain Tumor Detection Framework for MRI Analysis. Applied Sciences, 13, Article No. 3438. [Google Scholar] [CrossRef
[11] Gundogan, E. (2025) A Novel Hybrid Deep Learning Model Enhanced with Explainable AI for Brain Tumor Multi-Classification from MRI Images. Applied Sciences, 15, Article No. 5412. [Google Scholar] [CrossRef
[12] Srinivas, V.R. and Parvathi, R. (2026) Explainable AI-Driven MRI-Based Brain Tumor Classification: A Novel Deep Learning Approach. Frontiers in Artificial Intelligence, 8, Article ID: 1700214. [Google Scholar] [CrossRef
[13] Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D. and Batra, D. (2017) Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. 2017 IEEE International Conference on Computer Vision (ICCV), Venice, 22-29 October 2017, 618-626. [Google Scholar] [CrossRef
[14] Chattopadhyay, A., Sarkar, A., Howlader, P. and Balasubramanian, V.N. (2017) Grad-CAM++: Improved Visual Explanations for Deep Convolutional Networks.
[15] Ribeiro, M.T., Singh, S. and Guestrin, C. (2016) “Why Should I Trust You?”: Explaining the Predictions of Any Classifier. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, 13-17 August 2016, 1135-1144. [Google Scholar] [CrossRef
[16] Lundberg, S. and Lee, S.I. (2017) A Unified Approach to Interpreting Model Predictions.
[17] Zeiler, M.D. and Fergus, R. (2014) Visualizing and Understanding Convolutional Networks. In: Fleet, D., et al., Eds., Computer VisionECCV 2014, Springer International Publishing, 818-833. [Google Scholar] [CrossRef
[18] Iftikhar, S., Anjum, N., Siddiqui, A.B., Ur Rehman, M. and Ramzan, N. (2025) Explainable CNN for Brain Tumor Detection and Classification through XAI Based Key Features Identification. Brain Informatics, 12, Article No. 10. [Google Scholar] [CrossRef] [PubMed]
[19] Lecun, Y., Bottou, L., Bengio, Y. and Haffner, P. (1998) Gradient-Based Learning Applied to Document Recognition. Proceedings of the IEEE, 86, 2278-2324. [Google Scholar] [CrossRef
[20] Ning, Y., Liu, J., Qin, L., et al. (2023) A Novel Approach for Auto-Formulation of Optimization Problems.
[21] Rastogi, D., Johri, P., Donelli, M., Agarwal, T., Tiwari, S. and Singh, P. (2025) XAI-BT-EdgeNet: Explainable Edge-Aware Deep Learning with Squeeze-and-Excitation for Brain Tumor Detection and Prediction. Frontiers in Artificial Intelligence, 8, Article ID: 1676524. [Google Scholar] [CrossRef
[22] Singh, R., Gupta, S., Ibrahim, A.O., Gabralla, L.A., Bharany, S., Rehman, A.U., et al. (2025) Advanced Dynamic Ensemble Framework with Explainability Driven Insights for Precision Brain Tumor Classification across Datasets. Scientific Reports, 15, Article No. 29090. [Google Scholar] [CrossRef] [PubMed]
[23] Hafeez, Y., Memon, K., AL-Quraishi, M.S., Yahya, N., Elferik, S. and Ali, S.S.A. (2025) Explainable AI in Diagnostic Radiology for Neurological Disorders: A Systematic Review, and What Doctors Think about It. Diagnostics, 15, Article No. 168. [Google Scholar] [CrossRef] [PubMed]
[24] Yoon, H.C. and Lin, L.P. (2025) Brain Tumor Classification in MRI: Insights from LIME and Grad-CAM Explainable AI Techniques. IEEE Access, 13, 154172-154202. [Google Scholar] [CrossRef
[25] Gharaibeh, N. (2025) Enhancing Interpretability in Brain Tumor Detection: Leveraging Grad-CAM and SHAP for Explainable AI in MRI-Based Cancer Diagnosis. Applied Computer Science, 21, 182-197. [Google Scholar] [CrossRef
[26] Muhammad, D. and Bendechache, M. (2024) Unveiling the Black Box: A Systematic Review of Explainable Artificial Intelligence in Medical Image Analysis. Computational and Structural Biotechnology Journal, 24, 542-560. [Google Scholar] [CrossRef] [PubMed]
[27] Doshi-Velez, F. and Kim, B. (2017) Towards a Rigorous Science of Interpretable Machine Learning.
[28] Wang, D. and Tanaka, T. (2020) Robust Kernel Principal Component Analysis with ℓ2,1-Regularized Loss Minimization. IEEE Access, 8, 81864-81875. [Google Scholar] [CrossRef
[29] McKinney, S.M., Sieniek, M., Godbole, V., Godwin, J., Antropova, N., Ashrafian, H., et al. (2020) International Evaluation of an AI System for Breast Cancer Screening. Nature, 577, 89-94. [Google Scholar] [CrossRef] [PubMed]
[30] Böhle, M., Eitel, F., Weygandt, M. and Ritter, K. (2019) Layer-Wise Relevance Propagation for Explaining Deep Neural Network Decisions in MRI-Based Alzheimer’s Disease Classification. Frontiers in Aging Neuroscience, 11, Article No. 194. [Google Scholar] [CrossRef] [PubMed]