본문으로 건너뛰기
← 뒤로

Unsupervised SAM-guided mixture-of-multimodal-experts fusion network for medical image diagnosis.

Neural networks : the official journal of the International Neural Network Society 2026 Vol.196() p. 108348

Li J, Wu Y, Zheng X, Dong S

📝 환자 설명용 한 줄

Accurate diagnosis of cancer from medical images relies on both precise lesion localization and complementary multimodal information.

이 논문을 인용하기

BibTeX ↓ RIS ↓
APA Li J, Wu Y, et al. (2026). Unsupervised SAM-guided mixture-of-multimodal-experts fusion network for medical image diagnosis.. Neural networks : the official journal of the International Neural Network Society, 196, 108348. https://doi.org/10.1016/j.neunet.2025.108348
MLA Li J, et al.. "Unsupervised SAM-guided mixture-of-multimodal-experts fusion network for medical image diagnosis.." Neural networks : the official journal of the International Neural Network Society, vol. 196, 2026, pp. 108348.
PMID 41317632

Abstract

Accurate diagnosis of cancer from medical images relies on both precise lesion localization and complementary multimodal information. However, current methods suffer from two key limitations: (1) dependence on costly pixel-level annotations for lesion segmentation, and (2) rigid fusion strategies that ignore patient-specific modality contributions. To address these challenges, we propose an Unsupervised SAM-guided Mixture-of-Multimodal-Experts Fusion Network (UnSAM-MoME) for medical image diagnosis. In the first stage, we introduce a dual cross-validation segmentation network that automatically generates high-confidence prompts to guide the Segment Anything Model (SAM), enabling precise lesion localization without manual labels. In the second stage, we design a Mixture-of-Multimodal-Experts (MoME) fusion module that dynamically selects specialized experts to adaptively fuse image and metadata features based on individual patient characteristics. Experiments on three skin cancer datasets and one breast cancer dataset demonstrate that UnSAM-MoME achieves state-of-the-art performance, with significant improvements in accuracy, precision, and generalizability. Extensive ablation studies further validate the effectiveness of each module, underscoring the framework's potential for scalable and personalized cancer diagnosis.

MeSH Terms

Humans; Breast Neoplasms; Neural Networks, Computer; Skin Neoplasms; Unsupervised Machine Learning; Female; Image Interpretation, Computer-Assisted; Image Processing, Computer-Assisted

같은 제1저자의 인용 많은 논문 (5)