Convolutional neural networks for prostate cancer detection, classification, and segmentation: A systematic review and bibliometric analysis.
메타분석
2/5 보강
TL;DR
CNN-based prostate cancer detection, classification, and segmentation demonstrate promise for improving diagnostic accuracy and clinical workflow efficiency, though challenges in dataset standardization, regulatory compliance, and clinical integration remain to be addressed.
OpenAlex 토픽 ·
AI in cancer detection
Prostate Cancer Diagnosis and Treatment
Advanced Neural Network Applications
CNN-based prostate cancer detection, classification, and segmentation demonstrate promise for improving diagnostic accuracy and clinical workflow efficiency, though challenges in dataset standardizati
- 연구 설계 systematic review
APA
Burak Gülmez (2026). Convolutional neural networks for prostate cancer detection, classification, and segmentation: A systematic review and bibliometric analysis.. European journal of radiology open, 16, 100741. https://doi.org/10.1016/j.ejro.2026.100741
MLA
Burak Gülmez. "Convolutional neural networks for prostate cancer detection, classification, and segmentation: A systematic review and bibliometric analysis.." European journal of radiology open, vol. 16, 2026, pp. 100741.
PMID
41810091 ↗
Abstract 한글 요약
[BACKGROUND] Prostate cancer represents the second most common malignancy among men globally, necessitating accurate diagnostic methodologies for optimal patient outcomes. Convolutional neural networks (CNNs), a core deep learning methodology, have emerged as transformative technologies for automated prostate cancer detection, classification, and segmentation across multiple imaging modalities.
[MATERIALS AND METHODS] A systematic review following PRISMA guidelines was conducted across Web of Science, Scopus, and PubMed databases (January 2020-December 2025). CNN-based classification architectures were analyzed across ResNet, Vision Transformer, DenseNet, Xception, ConvNeXT, and Swin Transformer implementations, with comparative evaluation of accuracy and transfer learning performance. Object detection and segmentation approaches were examined across U-Net variants, R-CNN family algorithms, and YOLO-based implementations. Hyperparameter optimization strategies were assessed. Explainable AI methodologies including SHAP, Grad-CAM, DiCE, and LIME were evaluated for clinical interpretability and spatial localization accuracy.
[RESULTS] Analysis of 320 publications revealed peak research activity in 2024 (63 publications, 19.7%). The United States led with 58 publications (18.1%), followed by China with 55 (17.2%). Multiparametric MRI constituted the primary imaging modality (42.5%), followed by histopathology (28.1%), ultrasound (14.1%), and PET imaging (9.4%). Vision Transformer models demonstrated the highest classification accuracy among evaluated architectures, while U-Net variants dominated segmentation applications with consistently high Dice coefficients. SHAP emerged as the most frequently adopted explainability method across the reviewed studies.
[CONCLUSIONS] CNN-based prostate cancer detection, classification, and segmentation demonstrate promise for improving diagnostic accuracy and clinical workflow efficiency, though challenges in dataset standardization, regulatory compliance, and clinical integration remain to be addressed.
[MATERIALS AND METHODS] A systematic review following PRISMA guidelines was conducted across Web of Science, Scopus, and PubMed databases (January 2020-December 2025). CNN-based classification architectures were analyzed across ResNet, Vision Transformer, DenseNet, Xception, ConvNeXT, and Swin Transformer implementations, with comparative evaluation of accuracy and transfer learning performance. Object detection and segmentation approaches were examined across U-Net variants, R-CNN family algorithms, and YOLO-based implementations. Hyperparameter optimization strategies were assessed. Explainable AI methodologies including SHAP, Grad-CAM, DiCE, and LIME were evaluated for clinical interpretability and spatial localization accuracy.
[RESULTS] Analysis of 320 publications revealed peak research activity in 2024 (63 publications, 19.7%). The United States led with 58 publications (18.1%), followed by China with 55 (17.2%). Multiparametric MRI constituted the primary imaging modality (42.5%), followed by histopathology (28.1%), ultrasound (14.1%), and PET imaging (9.4%). Vision Transformer models demonstrated the highest classification accuracy among evaluated architectures, while U-Net variants dominated segmentation applications with consistently high Dice coefficients. SHAP emerged as the most frequently adopted explainability method across the reviewed studies.
[CONCLUSIONS] CNN-based prostate cancer detection, classification, and segmentation demonstrate promise for improving diagnostic accuracy and clinical workflow efficiency, though challenges in dataset standardization, regulatory compliance, and clinical integration remain to be addressed.
🏷️ 키워드 / MeSH 📖 같은 키워드 OA만
같은 제1저자의 인용 많은 논문 (1)
🏷️ 같은 키워드 · 무료전문 — 이 논문 MeSH/keyword 기반
- LCMS-Net: Deep Learning for Raw High Resolution Mass Spectrometry Data Applied to Forensic Cause-of-Death Screening.
- PIBAdb: a public cohort of multimodal colonoscopy videos and images including polyps with histological information.
- Exploring the Role of Extracellular Vesicles in Pancreatic and Hepatobiliary Cancers: Advances Through Artificial Intelligence.
- Automated RECOMIA AI-based total metabolic tumor volume in lymphoma - a retrospective study.
- Dual-stage pulmonary nodule detection in CT scans via cross-layer attention and adaptive multi-scale 3D CNN.
- Feasibility of Depth-in-Color En Face Optical Coherence Tomography for Colorectal Polyp Classification Using Ensemble Learning and Score-Level Fusion.