본문으로 건너뛰기
← 뒤로

Multimodal deep learning method based on multiple spectra for lung cancer early diagnosis.

1/5 보강
Spectrochimica acta. Part A, Molecular and biomolecular spectroscopy 📖 저널 OA 4.7% 2023: 0/1 OA 2024: 0/1 OA 2025: 0/13 OA 2026: 3/49 OA 2023~2026 2026 Vol.346() p. 126932
Retraction 확인
출처

Zhang H, Qi Y, Xu H, Lv R

📝 환자 설명용 한 줄

Lung cancer remains one of the most lethal malignancies worldwide, and the early and accurate diagnostic is critical.

이 논문을 인용하기

↓ .bib ↓ .ris
APA Zhang H, Qi Y, et al. (2026). Multimodal deep learning method based on multiple spectra for lung cancer early diagnosis.. Spectrochimica acta. Part A, Molecular and biomolecular spectroscopy, 346, 126932. https://doi.org/10.1016/j.saa.2025.126932
MLA Zhang H, et al.. "Multimodal deep learning method based on multiple spectra for lung cancer early diagnosis.." Spectrochimica acta. Part A, Molecular and biomolecular spectroscopy, vol. 346, 2026, pp. 126932.
PMID 40957203 ↗

Abstract

Lung cancer remains one of the most lethal malignancies worldwide, and the early and accurate diagnostic is critical. Traditional diagnostic techniques such as imaging and histopathology often suffer from limitations including high cost, radiation exposure, and reliance on expert experience. In this study, a multimodal deep learning method based on multiple spectra is proposed for lung cancer detection. The method integrates four common spectra including Fourier transform infrared spectra, UV-vis absorbance spectra, fluorescence spectra, and Raman spectra into a unified detection framework. Specifically, every spectrum sample is represented by its one-dimensional (1D) sequence and two-dimensional (2D) Gramian Angular Summation Field (GASF) image. A dual-branch architecture was designed to capture 1D patterns and 2D features. These complementary features are fused through a MambaVision-based fusion module to achieve efficient cross-modal interaction and global context modeling. The proposed method achieved outstanding performance with accuracy of 97.65 %, precision of 98.14 %, recall of 97.52 %, F1-score of 97.82 %, and AUC of 99.76 %. Moreover, the comparison and ablation experiments confirm the superiority of the training strategy and the proposed method. Interpretability analysis based on T-SNE and Class Activation Mapping proves that the model focuses on meaningful spectral regions in the detection process. This work demonstrates a promising paradigm for spectral-based intelligent diagnosis.

🏷️ 키워드 / MeSH 📖 같은 키워드 OA만

같은 제1저자의 인용 많은 논문 (5)

🏷️ 같은 키워드 · 무료전문 — 이 논문 MeSH/keyword 기반