Publication:
Can ChatGPT, an artificial intelligence language model, provide accurate and high-quality patient information on prostate cancer?

dc.contributor.authorCoşkun, Burhan
dc.contributor.authorOcakoğlu, Gökhan
dc.contributor.authorYetemen, Melih
dc.contributor.authorKaygısız, Onur
dc.contributor.buuauthorCOŞKUN, BURHAN
dc.contributor.buuauthorOCAKOĞLU, GÖKHAN
dc.contributor.buuauthorYETEMEN, MELİH
dc.contributor.buuauthorKAYGISIZ, ONUR
dc.contributor.departmentBursa Uludağ Üniversitesi/Tıp Fakültesi/Üroloji Anabilim Dalı.
dc.contributor.departmentBursa Uludağ Üniversitesi/Tıp Fakültesi/Biyoistatistik Anabilim Dalı.
dc.contributor.orcid0000-0002-8242-9921
dc.contributor.orcid0000-0002-1114-6051
dc.contributor.orcid0000-0002-9790-7295
dc.contributor.researcheridAAH-9704-2021
dc.contributor.researcheridL-9439-2019
dc.contributor.researcheridAAH-5180-2021
dc.contributor.researcheridJVM-2711-2024
dc.date.accessioned2024-10-02T08:16:47Z
dc.date.available2024-10-02T08:16:47Z
dc.date.issued2023-10-10
dc.description.abstractOBJECTIVE To evaluate the performance of ChatGPT, an artificial intelligence (AI) language model, in providing patient information on prostate cancer, and to compare the accuracy, similarity, and quality of the information to a reference source. METHODS Patient information material on prostate cancer was used as a reference source from the website of the European Association of Urology Patient Information. This was used to generate 59 queries. The accuracy of the model's content was determined with F1, precision, and recall scores. The similarity was assessed with cosine similarity, and the quality was evaluated using a 5RESULTS ChatGPT was able to respond to all prostate cancer-related queries. The average F1 score was 0.426 (range: 0-1), precision score was 0.349 (range: 0-1), recall score was 0.549 (range: 0-1), and cosine similarity was 0.609 (range: 0-1). The average GQS was 3.62 +/- 0.49 (range: 1-5), with no answers achieving the maximum GQS of 5. While ChatGPT produced a larger amount of information compared to the reference, the accuracy and quality of the content were not optimal, with all scores indicating need for improvement in the model's performance. CONCLUSION Caution should be exercised when using ChatGPT as a patient information source for prostate cancer due to limitations in its performance, which may lead to inaccuracies and potential misunderstandings. Further studies, using different topics and language models, are needed to fully understand the capabilities and limitations of AI-generated patient information. UROLOGY 180: 35-58, 2023. (c) 2023 Elsevier Inc. All rights reserved.
dc.identifier.doi10.1016/j.urology.2023.05.040
dc.identifier.endpage58
dc.identifier.issn0090-4295
dc.identifier.startpage35
dc.identifier.urihttps://doi.org/10.1016/j.urology.2023.05.040
dc.identifier.urihttps://www.sciencedirect.com/science/article/pii/S0090429523005708?via%3Dihub
dc.identifier.urihttps://hdl.handle.net/11452/45651
dc.identifier.volume180
dc.identifier.wos001127709200001
dc.indexed.wosWOS.SCI
dc.language.isoen
dc.publisherElsevier Science Inc
dc.relation.journalUrology
dc.relation.publicationcategoryMakale - Uluslararası Hakemli Dergi
dc.rightsinfo:eu-repo/semantics/closedAccess
dc.subjectHealth information
dc.subjectInternet
dc.subjectUrology & nephrology
dc.titleCan ChatGPT, an artificial intelligence language model, provide accurate and high-quality patient information on prostate cancer?
dc.typeArticle
dspace.entity.typePublication
relation.isAuthorOfPublication7e53dfda-90d9-48ee-acf2-c05fb2b33a29
relation.isAuthorOfPublication8ff963e8-284c-49e2-99b9-a46777690e8c
relation.isAuthorOfPublication28333c8d-6154-42ba-8784-f4f63d275cf4
relation.isAuthorOfPublicationc7166d91-f1a2-4296-b6cf-6666152225e0
relation.isAuthorOfPublication.latestForDiscovery7e53dfda-90d9-48ee-acf2-c05fb2b33a29

Files

Collections