Please use this identifier to cite or link to this item: https://hdl.handle.net/11499/58239
Full metadata record
DC FieldValueLanguage
dc.contributor.authorSağtaş, E.-
dc.contributor.authorUfuk, F.-
dc.contributor.authorPeker, H.-
dc.contributor.authorYağcı, A.B.-
dc.date.accessioned2024-11-20T18:03:35Z-
dc.date.available2024-11-20T18:03:35Z-
dc.date.issued2024-
dc.identifier.issn1309-9833-
dc.identifier.urihttps://doi.org/10.31362/patd.1487575-
dc.identifier.urihttps://search.trdizin.gov.tr/tr/yayin/detay/1275064-
dc.identifier.urihttps://hdl.handle.net/11499/58239-
dc.description.abstractPurpose: The advent of large language models like GPT-4 has opened new possibilities in natural language processing, with potential applications in medical literature. This study assesses GPT-4's ability to generate medical abstracts. It compares their quality to original abstracts written by human authors, aiming to understand the effectiveness of artificial intelligence in replicating complex, professional writing tasks. Materials and methods: A total of 250 original research articles from five prominent radiology journals published between 2021 and 2023 were selected. The body of these articles, excluding the abstracts, was fed into GPT-4, which then generated new abstracts. Three experienced radiologists blindly and independently evaluated all 500 abstracts using a five-point Likert scale for quality and understandability. Statistical analysis included mean score comparison inter-rater reliability using Fleiss' Kappa and Bland-Altman plots to assess agreement levels between raters. Results: Analysis revealed no significant difference in the mean scores between original and GPT-4 generated abstracts. The inter-rater reliability yielded kappa values indicating moderate to substantial agreement: 0.497 between Observers 1 and 2, 0.753 between Observers 1 and 3, and 0.645 between Observers 2 and 3. Bland-Altman analysis showed a slight systematic bias but was within acceptable limits of agreement. Conclusion: The study demonstrates that GPT-4 can generate medical abstracts with a quality comparable to those written by human experts. This suggests a promising role for artificial intelligence in facilitating the abstract writing process and improving its quality. © 2024, Pamukkale University. All rights reserved.en_US
dc.language.isoenen_US
dc.publisherPamukkale Universityen_US
dc.relation.ispartofPamukkale Medical Journalen_US
dc.rightsinfo:eu-repo/semantics/openAccessen_US
dc.subjectabstractsen_US
dc.subjectArtificial intelligenceen_US
dc.subjectChatGPTen_US
dc.subjectdiagnosisen_US
dc.subjectradiologyen_US
dc.titleArtificial intelligence meets medical expertise: evaluating GPT-4's proficiency in generating medical article abstractsen_US
dc.title.alternativeYapay zeka tıbbi uzmanlıkla buluşuyor: GPT-4'ün tıbbi makale özetleri oluşturmadaki yeterliliğinin değerlendirilmesien_US
dc.typeArticleen_US
dc.identifier.volume17en_US
dc.identifier.issue4en_US
dc.identifier.startpage756en_US
dc.identifier.endpage762en_US
dc.departmentPamukkale Universityen_US
dc.identifier.doi10.31362/patd.1487575-
dc.relation.publicationcategoryMakale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanıen_US
dc.authorscopusid6507716103-
dc.authorscopusid56600861000-
dc.authorscopusid57855999400-
dc.authorscopusid6507565069-
dc.identifier.scopus2-s2.0-85207505677en_US
dc.identifier.trdizinid1275064en_US
dc.institutionauthor-
item.cerifentitytypePublications-
item.openairetypeArticle-
item.openairecristypehttp://purl.org/coar/resource_type/c_18cf-
item.fulltextNo Fulltext-
item.languageiso639-1en-
item.grantfulltextnone-
crisitem.author.dept14.02. Internal Medicine-
crisitem.author.dept14.02. Internal Medicine-
crisitem.author.dept14.02. Internal Medicine-
Appears in Collections:Scopus İndeksli Yayınlar Koleksiyonu / Scopus Indexed Publications Collection
TR Dizin İndeksli Yayınlar Koleksiyonu / TR Dizin Indexed Publications Collection
Show simple item record



CORE Recommender

Google ScholarTM

Check




Altmetric


Items in GCRIS Repository are protected by copyright, with all rights reserved, unless otherwise indicated.