본문으로 건너뛰기
← 뒤로

Integration of deep learning and radiomic features from multiplex immunohistochemistry images for reproducible Multi-Outcome prediction in a Multi-Center study of colorectal cancer.

2/5 보강
International journal of medical informatics 📖 저널 OA 17.9% 2023: 1/1 OA 2024: 0/2 OA 2025: 0/3 OA 2026: 4/21 OA 2023~2026 2026 Vol.214() p. 106436 Radiomics and Machine Learning in Me
Retraction 확인
출처
PubMed DOI OpenAlex 마지막 보강 2026-04-28

PICO 자동 추출 (휴리스틱, conf 3/4)

유사 논문
P · Population 대상 환자/모집단
548 cases used for model training and internal testing, and 569 for external validation.
I · Intervention 중재 / 시술
multi-step selection (LASSO, MI, RFE) and were fused into a single feature space, followed by PCA-based dimensionality reduction
C · Comparison 대조 / 비교
추출되지 않음
O · Outcome 결과 / 결론
SHAP analysis confirmed feature interpretability, with fused features contributing the most across tasks. [CONCLUSIONS] This study demonstrates that fused mIHC-derived radiomic and deep features yield accurate, interpretable, and generalizable predictions for multiple CRC outcomes, supporting their integration into precision oncology workflows.
OpenAlex 토픽 · Radiomics and Machine Learning in Medical Imaging AI in cancer detection Ferroptosis and cancer prognosis

Yin Y, Sun Z, Deng X, Fan Q

📝 환자 설명용 한 줄

[OBJECTIVE] To develop and validate a robust, multimodal machine learning framework integrating radiomic and deep learning features from multiplex immunohistochemistry (mIHC) images for comprehensive

🔬 핵심 임상 통계 (초록에서 자동 추출 — 원문 검증 권장)
  • 표본수 (n) 71

이 논문을 인용하기

↓ .bib ↓ .ris
APA Yizhuo Yin, Zhe Sun, et al. (2026). Integration of deep learning and radiomic features from multiplex immunohistochemistry images for reproducible Multi-Outcome prediction in a Multi-Center study of colorectal cancer.. International journal of medical informatics, 214, 106436. https://doi.org/10.1016/j.ijmedinf.2026.106436
MLA Yizhuo Yin, et al.. "Integration of deep learning and radiomic features from multiplex immunohistochemistry images for reproducible Multi-Outcome prediction in a Multi-Center study of colorectal cancer.." International journal of medical informatics, vol. 214, 2026, pp. 106436.
PMID 41955913 ↗

Abstract

[OBJECTIVE] To develop and validate a robust, multimodal machine learning framework integrating radiomic and deep learning features from multiplex immunohistochemistry (mIHC) images for comprehensive outcome prediction in colorectal cancer (CRC).

[MATERIALS AND METHODS] This multi-institutional retrospective study included 2,117 CRC patients from seven centers, with 1,548 cases used for model training and internal testing, and 569 for external validation. mIHC-stained whole-slide images targeting six immune markers (CD3, CD8, CD45RO, PD-1, LAG-3, Tim-3) were analyzed from two spatial compartments: tumor center and invasive margin. Radiomic features (n = 71/region/marker) were extracted using HistomicsTK, while 768-dimensional deep features were derived using a pre-trained Vision Transformer (ViT-B/16). Feature robustness across biomarkers was quantified via intraclass correlation coefficients (ICC ≥ 0.75). Selected features underwent multi-step selection (LASSO, MI, RFE) and were fused into a single feature space, followed by PCA-based dimensionality reduction. Five clinical tasks were modeled: tumor recurrence, survival status, overall survival duration, TNM staging, and immune profile classification. Classification models (TabTransformer, XGBoost, TabNet) and survival models (DeepSurv, CoxPH, RSF) were trained using 5-fold cross-validation and tested on independent cohorts.

[RESULTS] Fused features significantly outperformed individual modalities across all tasks. TabTransformer with LASSO-selected fused features achieved top performance: recurrence (AUC = 95.9%), survival status (AUC = 94.5%), TNM staging (macro-AUC = 91.0%), and immune profile (macro-AUC = 91.0%). For survival regression, DeepSurv achieved a C-index of 0.82 and time-dependent AUC of 0.82. Models exhibited strong generalizability, with negligible performance drop on external datasets. SHAP analysis confirmed feature interpretability, with fused features contributing the most across tasks.

[CONCLUSIONS] This study demonstrates that fused mIHC-derived radiomic and deep features yield accurate, interpretable, and generalizable predictions for multiple CRC outcomes, supporting their integration into precision oncology workflows.

🏷️ 키워드 / MeSH 📖 같은 키워드 OA만

같은 제1저자의 인용 많은 논문 (5)

🏷️ 같은 키워드 · 무료전문 — 이 논문 MeSH/keyword 기반