Publication:
Evaluating the Performance of ChatGPT in Urology: A Comparative Study of Knowledge Interpretation and Patient Guidance.

dc.contributor.authorŞAHİN, BAHADIR
dc.contributor.authorGENÇ, YUNUS EMRE
dc.contributor.authorsŞahin B., Genç Y. E., Doğan K., Şener T. E., Şekerci Ç. A., Tanıdır Y., Yücel S., Tarcan T., Çam H. K.
dc.date.accessioned2024-06-03T10:26:08Z
dc.date.accessioned2026-01-11T15:34:23Z
dc.date.available2024-06-03T10:26:08Z
dc.date.issued2024-03-01
dc.description.abstractBackground/Aim: To evaluate the performance of Chat Generative Pre-trained Transformer (ChatGPT), a large language model trained by Open artificial intelligence. Materials and Methods: This study has three main steps to evaluate the effectiveness of ChatGPT in the urologic field. The first step involved 35 questions from our institution's experts, who have at least 10 years of experience in their fields. The responses of ChatGPT versions were qualitatively compared with the responses of urology residents to the same questions. The second step assesses the reliability of ChatGPT versions in answering current debate topics. The third step was to assess the reliability of ChatGPT versions in providing medical recommendations and directives to patients' commonly asked questions during the outpatient and inpatient clinic. Results: In the first step, version 4 provided correct answers to 25 questions out of 35 while version 3.5 provided only 19 (71.4% vs 54%). It was observed that residents in their last year of education in our clinic also provided a mean of 25 correct answers, and 4th year residents provided a mean of 19.3 correct responses. The second step involved evaluating the response of both versions to debate situations in urology, and it was found that both versions provided variable and inappropriate results. In the last step, both versions had a similar success rate in providing recommendations and guidance to patients based on expert ratings. Conclusion: The difference between the two versions of the 35 questions in the first step of the study was thought to be due to the improvement of ChatGPT's literature and data synthesis abilities. It may be a logical approach to use ChatGPT versions to inform the nonhealth care providers' questions with quick and safe answers but should not be used to as a diagnostic tool or make a choice among different treatment modalities.
dc.identifier.citationŞahin B., Genç Y. E., Doğan K., Şener T. E., Şekerci Ç. A., Tanıdır Y., Yücel S., Tarcan T., Çam H. K., "Evaluating the Performance of ChatGPT in Urology: A Comparative Study of Knowledge Interpretation and Patient Guidance. ", JOURNAL OF ENDOUROLOGY, cilt.28, sa.3, ss.1-10, 2024
dc.identifier.doi10.1089/end.2023.0413
dc.identifier.endpage10
dc.identifier.issn0892-7790
dc.identifier.issue3
dc.identifier.startpage1
dc.identifier.urihttps://hdl.handle.net/11424/296958
dc.identifier.volume28
dc.language.isoeng
dc.relation.ispartofJOURNAL OF ENDOUROLOGY
dc.rightsinfo:eu-repo/semantics/openAccess
dc.subjectTıp
dc.subjectCerrahi Tıp Bilimleri
dc.subjectÜroloji
dc.subjectSağlık Bilimleri
dc.subjectMedicine
dc.subjectSurgery Medicine Sciences
dc.subjectUrology
dc.subjectHealth Sciences
dc.subjectKlinik Tıp (MED)
dc.subjectKlinik Tıp
dc.subjectÜROLOJİ VE NEFROLOJİ
dc.subjectCERRAHİ
dc.subjectClinical Medicine (MED)
dc.subjectCLINICAL MEDICINE
dc.subjectUROLOGY & NEPHROLOGY
dc.subjectSURGERY
dc.subjectCerrahi
dc.subjectNefroloji
dc.subjectSurgery
dc.subjectNephrology
dc.titleEvaluating the Performance of ChatGPT in Urology: A Comparative Study of Knowledge Interpretation and Patient Guidance.
dc.typearticle
dspace.entity.typePublication

Files