Пожалуйста, используйте этот идентификатор, чтобы цитировать или ссылаться на этот ресурс: https://dspace.uzhnu.edu.ua/jspui/handle/lib/68102
Полная запись метаданных
Поле DCЗначениеЯзык
dc.contributor.authorМихалко, Ярослав Омелянович-
dc.contributor.authorФілак, Ярослав Феліксович-
dc.contributor.authorДуткевич-Іванська, Юлія Василівна-
dc.contributor.authorСабадош, Мар’яна Володимирівна-
dc.contributor.authorРубцова, Єлізавета Іллівна-
dc.date.accessioned2024-11-30T09:34:42Z-
dc.date.available2024-11-30T09:34:42Z-
dc.date.issued2024-10-
dc.identifier.citationFrom open-ended to multiple-choice: evaluating diagnostic performance and consistency of ChatGPT, Google Gemini and Claude AI / Y. O. Mykhalko, Y. F. Filak, Y. V. Dutkevych-Ivanska, M. V. Sabadosh, Y. I. Rubtsova // Wiadomości Lekarskie Medical Advances. – 2024, – Vol. 77(10). – p. 1852-1856.uk
dc.identifier.issn0043-5147-
dc.identifier.urihttps://dspace.uzhnu.edu.ua/jspui/handle/lib/68102-
dc.description.abstractAim: To determine the performance and response repeatability of freely available LLMs in diagnosing diseases based on clinical case descriptions. Materials and Methods: 100 detailed clinical case descriptions were used to evaluate the diagnostic performance of ChatGPT 3.5, ChatGPT 4o, Google Gemini, and Claude AI 3.5 Sonnet large language models (LLMs). The analysis was conducted in two phases: Phase 1 with only case descriptions, and Phase 2 with descriptions and answer variants. Each phase used specific prompts and was repeated twice to assess agreement. Response consistency was determined using agreement percentage and Cohen's Kappa (k). 95% confidence intervals for proportions were calculated using Wilson's method. Statistical significance was set at p<0.05 using Fisher's exact test. Results: In Phase 1 of the study, ChatGPT 3.5, ChatGPT 4o, Google Gemini, and Claude AI 3.5 Sonnet's efficacy was 69.00%, 64.00%, 44.00%, and 72.00% respectively. All models showed high consistency as agreement percentages ranged from 93.00% to 97.00%, and k ranged from 0.86 to 0.94. In Phase 2 all models' productivity increased significantly (90.00%, 95.00%, 65.00%, and 89.00% for ChatGPT 3.5, ChatGPT 4o, Google Gemini, and Claude AI 3.5 Sonnet respectively). The agreement percentages ranged from 97.00% to 99.00%, while k values were between 0.85 and 0.93. Conclusion: Claude AI 3.5 Sonnet and both ChatGPT models can be used effectively for the differential diagnosis process, while using these models for diagnosing from scratch should be done with caution. As Google Gemini's efficacy was low, its feasibility in real clinical practice is currently questionable.uk
dc.language.isoenuk
dc.publisherALUNA Publishinguk
dc.subjectartificial intelligenceuk
dc.subjectlarge language modeluk
dc.subjectdiagnosisuk
dc.subjectperformanceuk
dc.titleFrom Open-Ended to Multiple-Choice: Evaluating Diagnostic Performance and Consistency of ChatGPT, Google Gemini and Claude AIuk
dc.title.alternativeFrom Open-Ended to Multiple-Choice: Evaluating Diagnostic Performance and Consistency of ChatGPT, Google Gemini and Claude AIuk
dc.typeTextuk
dc.pubTypeСтаттяuk
Располагается в коллекциях:Наукові публікації кафедри терапії та сімейної медицини

Файлы этого ресурса:
Файл Описание РазмерФормат 
article-wiadomosci-2024.pdf3.46 MBAdobe PDFПросмотреть/Открыть


Все ресурсы в архиве электронных ресурсов защищены авторским правом, все права сохранены.