Explainable deep learning-based lung cancer diagnosis using clinically-guided local interpretable model-agnostic explanations.
2/5 보강
OpenAlex 토픽 ·
Lung Cancer Diagnosis and Treatment
COVID-19 diagnosis using AI
Explainable Artificial Intelligence (XAI)
Lung cancer remains one of the leading causes of cancer-related deaths worldwide, highlighting the urgent need for accurate and interpretable diagnostic tools.
- Specificity 99.74%
APA
Shahab Ul Hassan, Said Jadid Abdulkadir, et al. (2026). Explainable deep learning-based lung cancer diagnosis using clinically-guided local interpretable model-agnostic explanations.. Scientific reports. https://doi.org/10.1038/s41598-026-44127-x
MLA
Shahab Ul Hassan, et al.. "Explainable deep learning-based lung cancer diagnosis using clinically-guided local interpretable model-agnostic explanations.." Scientific reports, 2026.
PMID
41957146 ↗
Abstract 한글 요약
Lung cancer remains one of the leading causes of cancer-related deaths worldwide, highlighting the urgent need for accurate and interpretable diagnostic tools. While deep learning (DL) models have achieved strong results in medical image classification, their opaque decision-making process remains a barrier to clinical adoption. This study proposes an adaptive superpixel perturbation-based local interpretable model-agnostic explanations (ASP-LIME), a novel explanation framework designed to generate faithful and localized interpretations of DL predictions, providing insights into the model's decision-making process. The proposed approach improves upon the original local interpretable model-agnostic explanations method by introducing adaptive superpixel segmentation, stratified perturbation strategies, lung region masking, and post-processing enhancements tailored for medical imaging. The proposed framework is applied to a lung cancer classification task using a custom-designed convolutional neural network, MedDeepNet, as the predictive model. Experimental results on a publicly available lung image dataset demonstrate that MedDeepNet achieves 99.84% accuracy, 99.66% recall, 99.82% precision, 99.74% specificity, and a 99.74% F1-score. ASP-LIME produces high-fidelity explanations with strong localization to pathological regions, achieving scores of 0.0300 for deletion, 0.9622 for insertion, and 0.9661 for Area Between Perturbation Curves (ABPC), surpassing typical benchmarks for interpretability methods. The findings demonstrate that the proposed framework offers consistent and interpretable explanations that enhance understanding of model decisions in medical imaging applications.
🏷️ 키워드 / MeSH 📖 같은 키워드 OA만
🏷️ 같은 키워드 · 무료전문 — 이 논문 MeSH/keyword 기반
- LCMS-Net: Deep Learning for Raw High Resolution Mass Spectrometry Data Applied to Forensic Cause-of-Death Screening.
- PIBAdb: a public cohort of multimodal colonoscopy videos and images including polyps with histological information.
- Exploring the Role of Extracellular Vesicles in Pancreatic and Hepatobiliary Cancers: Advances Through Artificial Intelligence.
- Feasibility of Depth-in-Color En Face Optical Coherence Tomography for Colorectal Polyp Classification Using Ensemble Learning and Score-Level Fusion.
- Impact of CT Intensity and Contrast Variability on Deep-Learning-Based Lung-Nodule Detection: A Systematic Review of Preprocessing and Harmonization Strategies (2020-2025).
- A Transformer-Based Deep Learning Model for predicting Early Recurrence in Hepatocellular Carcinoma After Hepatectomy Using Intravoxel Incoherent Motion Images.