Rethinking Multi-center Semi-supervised Breast Cancer Ultrasound Image Segmentation: An Intermediate-domain Perspective.
1/5 보강
Multi-center breast ultrasound images eg mentation aims to leverage limited labeled data from a single center to enhance model discriminability across unlabeled data from other centers.
APA
Ye Z, Zhang Y, et al. (2026). Rethinking Multi-center Semi-supervised Breast Cancer Ultrasound Image Segmentation: An Intermediate-domain Perspective.. IEEE journal of biomedical and health informatics, PP. https://doi.org/10.1109/JBHI.2026.3668217
MLA
Ye Z, et al.. "Rethinking Multi-center Semi-supervised Breast Cancer Ultrasound Image Segmentation: An Intermediate-domain Perspective.." IEEE journal of biomedical and health informatics, vol. PP, 2026.
PMID
41740111
Abstract
Multi-center breast ultrasound images eg mentation aims to leverage limited labeled data from a single center to enhance model discriminability across unlabeled data from other centers. However, differences in equipment parameters, disease severity, and imaging conditions collectively contribute to significant cross domain shifts in multi-center data. In a spirit of the golden mean, we argue that constructing an intermediate domain between the source and target domains can effectively improve model generalization. Therefore, we propose a Cross-domain Few-label Generalization (CFG) framework for multi-center breast ultrasound image segmentation. Specifically, we design the Intermediate Domain Generator (IDG) to generate intermediate domain samplesthatcontain features from both the source and target domains bidirectionally, enabling the model to explicitly learn univer sal semantic representations. Additionally, we apply Swin Masked Autoencoder (MAE) to mask and reconstruct ul trasound images, simulating speckle noise encountered during clinical ultrasound acquisition, thereby increasing the diversity of intermediate domain samples. Further more, we integrate the Kolmogorov-Arnold Network (KAN) with UNet to construct KAN-UNet, integrating learnable spline functions directly onto the edges, enabling effective multi-scale perception of breast cancer lesion features. Experimental results show that even with limited labeled data from the source domain (BUSI-WHU), the CFG frame work achieves a Kappa value of 77.17%, surpassing ten state-of-the-art methods and outperforming the second best method by 0.78% across four multi-center ultrasound datasets (BUSI-WHU, BUSI, Dataset-B, and Dataset-C) collected from different medical centers. The code is available at https://github.com/yzygit1230/CFG.
같은 제1저자의 인용 많은 논문 (5)
- Decoding the mitochondrial metabolic engine for tumor metastasis and clinical therapeutic opportunities.
- Single-Cell Transcriptomic Landscape of Right-Sided Colon Cancer Reveals Cellular and Molecular Features of Metastatic Potential.
- Nanomedicine strategies for disulfidptosis activation in SLC7A11-high tumors.
- Genome-wide analysis highlights epigenetic link to age at menarche.
- HAMIL: Hierarchical Attention Multi-Instance Learning for Label-Free Colorectal Cancer Typing.