본문으로 건너뛰기
← 뒤로

Man Versus Machine: A Comparative Study of Human and ChatGPT-Generated Abstracts in Plastic Surgery Research.

Aesthetic plastic surgery 2025 Vol.49(17) p. 5013-5020 🌐 cited 8 Artificial Intelligence in Healthcar
TL;DR ChatGPT holds promise in expediting the creation of high-quality scientific abstracts, potentially enhancing efficiency in research and scientific writing tasks, however, due to its exploratory nature, this study calls for additional research to validate these promising findings.
📈 연도별 인용 (2025–2026) · 합계 8
OpenAlex 토픽 · Artificial Intelligence in Healthcare and Education Radiomics and Machine Learning in Medical Imaging Anatomy and Medical Technology

Pressman SM, Garcia JP, Borna S, Gomez-Cabello CA, Haider SA, Haider CR, Forte AJ

📝 환자 설명용 한 줄

【연구 목적】 ChatGPT-4가 성형외과 연구 논문의 초록을 작성하는 데 있어 인간 저자와 비교하여 얼마나 효과적이고 고품질의 텍스트를 생성할 수 있는지 평가하는 것을 목적으로 한다.

이 논문을 인용하기

BibTeX ↓ RIS ↓
APA Sophia M. Pressman, John P. Garcia, et al. (2025). Man Versus Machine: A Comparative Study of Human and ChatGPT-Generated Abstracts in Plastic Surgery Research.. Aesthetic plastic surgery, 49(17), 5013-5020. https://doi.org/10.1007/s00266-025-04836-6
MLA Sophia M. Pressman, et al.. "Man Versus Machine: A Comparative Study of Human and ChatGPT-Generated Abstracts in Plastic Surgery Research.." Aesthetic plastic surgery, vol. 49, no. 17, 2025, pp. 5013-5020.
PMID 40229613

Abstract

[BACKGROUND] Since its 2022 release, ChatGPT has gained recognition for its potential to expedite time-consuming writing tasks like scientific writing. Well-written scientific abstracts are essential for clear and efficient communication of research findings. This study aims to explore ChatGPT-4's capability to produce well-crafted abstracts.

[METHODS] Ten abstract-less plastic surgery articles from PubMed were uploaded to ChatGPT, each with a prompt to generate one abstract. Flesch-Kincaid Grade Level (FKGL) and Flesch Reading Ease Score (FRES) were calculated for all abstracts. Additionally, three physician evaluators blindly assessed the ten original and ten ChatGPT-generated abstracts using a 5-point Likert scale. Results were compared and analyzed using descriptive statistics with mean and standard deviation (SD).

[RESULTS] The original abstracts averaged an FKGL of 14.1 (SD 2.9) and an FRES of 25.2 (SD 14.2), while ChatGPT-generated abstracts had scores of 15.6 (SD 2.4) and 15.4 (SD 13.1), respectively. Collectively, evaluators identified two-thirds of the ChatGPT abstracts, but preferred the ChatGPT abstracts 90% of the time. On average, the evaluators found the ChatGPT abstracts to be more "well written" (4.23 vs. 3.50, p value < 0.001) and "clear and concise" (4.30 vs. 3.53, p value < 0.001) compared to the original abstracts.

[CONCLUSIONS] Despite a slightly higher reading level, evaluators generally preferred ChatGPT abstracts, which received higher ratings overall. These findings suggest ChatGPT holds promise in expediting the creation of high-quality scientific abstracts, potentially enhancing efficiency in research and scientific writing tasks. However, due to its exploratory nature, this study calls for additional research to validate these promising findings.

[LEVEL OF EVIDENCE IV] This journal requires that authors assign a level of evidence to each article. For a full description of these Evidence-Based Medicine ratings, please refer to the Table of Contents or the online Instructions to Authors   www.springer.com/00266.

추출된 의학 개체 (NER)

유형영어 표현한국어 / 풀이UMLS CUI출처등장
약물 ChatGPT scispacy 1
기타 Human scispacy 1

MeSH Terms

Humans; Surgery, Plastic; Abstracting and Indexing; Biomedical Research; Periodicals as Topic; Generative Artificial Intelligence