Fully automated 3D multi-modal deep learning model for preoperative T-stage prediction of colorectal cancer using F-FDG PET/CT.
[PURPOSE] This study aimed to develop a fully automated 3D multi-modal deep learning model using preoperative F-FDG PET/CT to predict the T-stage of colorectal cancer (CRC) and evaluate its clinical u
- p-value P < 0.05
APA
Zhang M, Li Y, et al. (2026). Fully automated 3D multi-modal deep learning model for preoperative T-stage prediction of colorectal cancer using F-FDG PET/CT.. European journal of nuclear medicine and molecular imaging, 53(2), 910-920. https://doi.org/10.1007/s00259-025-07450-5
MLA
Zhang M, et al.. "Fully automated 3D multi-modal deep learning model for preoperative T-stage prediction of colorectal cancer using F-FDG PET/CT.." European journal of nuclear medicine and molecular imaging, vol. 53, no. 2, 2026, pp. 910-920.
PMID
40719866
Abstract
[PURPOSE] This study aimed to develop a fully automated 3D multi-modal deep learning model using preoperative F-FDG PET/CT to predict the T-stage of colorectal cancer (CRC) and evaluate its clinical utility.
[METHODS] A retrospective cohort of 474 CRC patients was included, with 400 patients for internal cohort and 74 patients for external cohort. Patients were classified into early T-stage (T1-T2) and advanced T-stage (T3-T4) groups. Automatic segmentation of the volume of interest (VOI) was achieved based on TotalSegmentator. A 3D ResNet18-based deep learning model integrated with a cross-multi-head attention mechanism was developed. Five models (CT + PET + Clinic (CPC), CT + PET (CP), PET (P), CT (C), Clinic) and two radiologists' assessment were compared. Performance was evaluated using Area Under the Curve (AUC). Grad-CAM was employed to provide visual interpretability of decision-critical regions.
[RESULTS] The automated segmentation achieved Dice scores of 0.884 (CT) and 0.888 (PET). The CPC and CP models achieved superior performance, with AUCs of 0.869 and 0.869 in the internal validation cohort, respectively, outperforming single-modality models (P: 0.832; C: 0.809; Clinic: 0.728) and the radiologists (AUC: 0.627, P < 0.05 for all models vs. radiologists, except for the Clinical model). External validation exhibited a similar trend, with AUCs of 0.814, 0.812, 0.763, 0.714, 0.663 and 0.704, respectively. Grad-CAM visualization highlighted tumor-centric regions for early T-stage and peri-tumoral tissue infiltration for advanced T-stage.
[CONCLUSION] The fully automated multimodal, fusing PET/CT with cross-multi-head-attention, improved T-stage prediction in CRC, surpassing the single-modality models and radiologists, offering a time-efficient tool to aid clinical decision-making.
[METHODS] A retrospective cohort of 474 CRC patients was included, with 400 patients for internal cohort and 74 patients for external cohort. Patients were classified into early T-stage (T1-T2) and advanced T-stage (T3-T4) groups. Automatic segmentation of the volume of interest (VOI) was achieved based on TotalSegmentator. A 3D ResNet18-based deep learning model integrated with a cross-multi-head attention mechanism was developed. Five models (CT + PET + Clinic (CPC), CT + PET (CP), PET (P), CT (C), Clinic) and two radiologists' assessment were compared. Performance was evaluated using Area Under the Curve (AUC). Grad-CAM was employed to provide visual interpretability of decision-critical regions.
[RESULTS] The automated segmentation achieved Dice scores of 0.884 (CT) and 0.888 (PET). The CPC and CP models achieved superior performance, with AUCs of 0.869 and 0.869 in the internal validation cohort, respectively, outperforming single-modality models (P: 0.832; C: 0.809; Clinic: 0.728) and the radiologists (AUC: 0.627, P < 0.05 for all models vs. radiologists, except for the Clinical model). External validation exhibited a similar trend, with AUCs of 0.814, 0.812, 0.763, 0.714, 0.663 and 0.704, respectively. Grad-CAM visualization highlighted tumor-centric regions for early T-stage and peri-tumoral tissue infiltration for advanced T-stage.
[CONCLUSION] The fully automated multimodal, fusing PET/CT with cross-multi-head-attention, improved T-stage prediction in CRC, surpassing the single-modality models and radiologists, offering a time-efficient tool to aid clinical decision-making.
MeSH Terms
Humans; Positron Emission Tomography Computed Tomography; Deep Learning; Fluorodeoxyglucose F18; Colorectal Neoplasms; Male; Female; Middle Aged; Retrospective Studies; Aged; Neoplasm Staging; Imaging, Three-Dimensional; Automation; Preoperative Period; Adult
같은 제1저자의 인용 많은 논문 (5)
- Deep Learning-based Multiple Arterial Phase MRI: A Step toward Improved Hepatocellular Carcinoma Diagnosis.
- Artificial intelligence-driven pharmacovigilance analysis of polypharmacy and adverse events in breast cancer survivors treated with antidepressants.
- Explainable artificial intelligence for multi-modal cancer analysis: From genomics to immunology.
- pH and temperature dual-responsive covalent organic framework composites for efficient enrichment of phosphoproteins.
- Novel nanoparticle-combined therapy through EGFR and PD-L1 blockade: PFPR-siRNA-mediated suppressing NSCLC proliferation via PDT and enhanced antitumor immunity.