Comparative Analysis of Large Language Models in Simplifying Turkish Ultrasound Reports to Enhance Patient Understanding
dc.authorid | Camur, Eren/0000-0002-8774-5800 | |
dc.authorid | Cesur, Turay/0000-0002-2726-8045 | |
dc.contributor.author | Gunes, Yasin Celal | |
dc.contributor.author | Cesur, Turay | |
dc.contributor.author | Camur, Eren | |
dc.date.accessioned | 2025-01-21T16:36:41Z | |
dc.date.available | 2025-01-21T16:36:41Z | |
dc.date.issued | 2024 | |
dc.department | Kırıkkale Üniversitesi | |
dc.description.abstract | Objective: To evaluate and compare the abilities of Language Models (LLMs) in simplifying Turkish ultrasound (US) findings for patients. Methods: We assessed the simplification performance of four LLMs: ChatGPT 4, Gemini 1.5 Pro, Claude 3 Opus, and Perplexity, using fifty fictional Turkish US findings. Comparison was based on Ate man's Readability Index and word count. Three radiologists rated medical accuracy, consistency, and comprehensibility on a Likert scale from 1 to 5. Statistical tests (Friedman, Wilcoxon, and Spearman correlation) examined differences in LLMs' performance. Results: Gemini 1.5 Pro, ChatGPT-4, and Claude 3 Opus received high Likert scores for medical accuracy, consistency, and comprehensibility (mean: 4.7-4.8). Perplexity scored significantly lower (mean: 4.1, p<0.001). Gemini 1.5 Pro achieved the highest readability score (mean: 61.16), followed by ChatGPT-4 (mean: 58.94) and Claude 3 Opus (mean: 51.16). Perplexity had the lowest readability score (mean: 47.01). Gemini 1.5 Pro and ChatGPT-4 used significantly more words compared to Claude 3 Opus and Perplexity (p<0.001). Linear correlation analysis revealed a positive correlation between word count of fictional US findings and responses generated by Gemini 1.5 Pro (correlation coefficient = 0.38, p<0.05) and ChatGPT-4 (correlation coefficient = 0.43, p<0.001). Conclusion: This study highlights strong potential of LLMs in simplifying Turkish US and Claude 3 Opus performed well, highlighting their effectiveness in healthcare communication. Further research is required to fully understand the integration of making. | |
dc.description.sponsorship | Funding : No funding was received for this study. | |
dc.identifier.doi | 10.58600/eurjther2225 | |
dc.identifier.issn | 2564-7784 | |
dc.identifier.issn | 2564-7040 | |
dc.identifier.uri | https://doi.org/10.58600/eurjther2225 | |
dc.identifier.uri | https://hdl.handle.net/20.500.12587/24366 | |
dc.identifier.wos | WOS:001289504400001 | |
dc.identifier.wosquality | N/A | |
dc.indekslendigikaynak | Web of Science | |
dc.language.iso | en | |
dc.publisher | Pera Yayincilik Hizmetleri | |
dc.relation.ispartof | European Journal of Therapeutics | |
dc.relation.publicationcategory | Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı | |
dc.rights | info:eu-repo/semantics/openAccess | |
dc.snmz | KA_20241229 | |
dc.subject | Large Language Models; ChatGPT; Claude 3 Opus; Ultrasound; Simplify | |
dc.title | Comparative Analysis of Large Language Models in Simplifying Turkish Ultrasound Reports to Enhance Patient Understanding | |
dc.type | Article |