본문으로 건너뛰기
← 뒤로

Fine-tuned large language models with structured prompts enable efficient construction of lung cancer knowledge graphs.

Scientific reports 2026 Vol.16(1)

Zhou C, Gong Q, Luan H, Zhan W, Zhu J, Zhang Q

📝 환자 설명용 한 줄

[UNLABELLED] Conventional methods for constructing lung cancer knowledge graphs require extensive annotated data, resulting in high construction costs.

이 논문을 인용하기

BibTeX ↓ RIS ↓
APA Zhou C, Gong Q, et al. (2026). Fine-tuned large language models with structured prompts enable efficient construction of lung cancer knowledge graphs.. Scientific reports, 16(1). https://doi.org/10.1038/s41598-026-38959-w
MLA Zhou C, et al.. "Fine-tuned large language models with structured prompts enable efficient construction of lung cancer knowledge graphs.." Scientific reports, vol. 16, no. 1, 2026.
PMID 41708685

Abstract

[UNLABELLED] Conventional methods for constructing lung cancer knowledge graphs require extensive annotated data, resulting in high construction costs. To address this challenge, this study developed the Knowledge Graph Large Model (KGLM) through a fine-tuning strategy to efficiently extract lung cancer knowledge triples. Carefully designed prompts were used during knowledge extraction, efficiently process complex, unstructured lung cancer information. Simultaneously, semi-structured clinical data was integrated with structured public graph data, and an entity alignment approach based on Jaccard similarity and Sentence-BERT (SBERT) successfully constructed the Lung Cancer Knowledge Graph (LCKG). The experimental outcomes highlighted the significance of our unified framework, which integrates prompt engineering and fine-tuning. Notably, the KGLM model with its structured prompts demonstrated superior performance in relation extraction tasks on large datasets, achieving an F1 score of 82%, a 25% improvement over baseline models. Furthermore, comparisons with traditional deep learning methods validated the effectiveness and suitability of employing large language models for knowledge graph construction.

[SUPPLEMENTARY INFORMATION] The online version contains supplementary material available at 10.1038/s41598-026-38959-w.

같은 제1저자의 인용 많은 논문 (5)