본문으로 건너뛰기
← 뒤로

Ensemble Deep Learning-Based High-Precision Framework for Breast Cancer Detection from Histopathological Images.

Diagnostics (Basel, Switzerland) 2026 Vol.16(5)

Ahmad F, Jaffar A, Latif G, Alghazo J, Bhatti SM

📝 환자 설명용 한 줄

: Analysis of histopathological images is the absolute standard of breast cancer diagnosis.

이 논문을 인용하기

BibTeX ↓ RIS ↓
APA Ahmad F, Jaffar A, et al. (2026). Ensemble Deep Learning-Based High-Precision Framework for Breast Cancer Detection from Histopathological Images.. Diagnostics (Basel, Switzerland), 16(5). https://doi.org/10.3390/diagnostics16050653
MLA Ahmad F, et al.. "Ensemble Deep Learning-Based High-Precision Framework for Breast Cancer Detection from Histopathological Images.." Diagnostics (Basel, Switzerland), vol. 16, no. 5, 2026.
PMID 41827929

Abstract

: Analysis of histopathological images is the absolute standard of breast cancer diagnosis. However, modern deep learning- and ViT-based architecture still struggle to capture effective local and global discriminatory patterns that tend to make architecture more complex, increasing the risk of overfitting and optimization problems. : To address these problems, this paper proposes a four-phase hybrid framework that aims to enhance the feature fusion, improving the model's strength, robustness, and generalization ability. In Phase 1, the BreakHis dataset was split patient-wise into a 70-15-15 manner to avoid data leakage, while extensive data augmentation, comprehensive normalization, and a five-fold cross-validation protocol were implemented to make the dataset more varied and reliably evaluated without bias. Phase 2 entailed the training of three CNNs (VGG16, ResNet50, and DenseNet121) and four ViTs (DeiT, CaiT, T2T-ViT, and Swin Transformer) independently to establish the strict baseline performance standards. In Phase 3, the CNN-based features were fused and classified with a soft voting mechanism to allow more stable and representative learning. Phase 4 depicts the Proposed Framework, which combines the two best-performing CNN and ViT models. Feature refinements were performed randomly by using Global Average Pooling and feature scaling, while a self-attention mechanism enabled the accurate cross-modal feature fusion. The generalization capability of the fused representation was further enhanced by the subsequent of dense layers followed by dropout. : XGBoost exhibited the highest performance among the evaluated ML classifiers, achieving 98.7% accuracy and 98.7% F1-score on BreakHis, while achieving 95.8% accuracy on external BACH dataset backed by Grad-CAM- and Grad-CAM++-based interpretability. : By integrating CNNs and ViTs through self-attention, the proposed framework offers a robust and interpretable solution for automated breast cancer diagnosis.

같은 제1저자의 인용 많은 논문 (3)