This is a preview of subscription content, access via your institution
Relevant articles
Open Access articles citing this article.
-
Comparison of the readability of ChatGPT and Bard in medical communication: a meta-analysis
BMC Medical Informatics and Decision Making Open Access 01 September 2025
Access options
Subscribe to this journal
Receive 18 print issues and online access
$259.00 per year
only $14.39 per issue
Buy this article
- Purchase on SpringerLink
- Instant access to full article PDF
Prices may be subject to local taxes which are calculated during checkout
References
Zhao FF, He HJ, Liang JJ, Cen J, Wang Y, Lin H, et al. Benchmarking the performance of large language models in uveitis: a comparative analysis of ChatGPT-3.5, ChatGPT-4.0, Google Gemini, and Anthropic Claude3. Eye. 2024. https://doi.org/10.1038/s41433-024-03545-9.
Lim ZW, Pushpanathan K, Yew SME, Lai Y, Sun CH, Lam JSH, et al. Benchmarking large language models’ performances for myopia care: a comparative analysis of ChatGPT-3.5, ChatGPT-4.0, and Google Bard. EBioMedicine. 2023;95:104770. https://doi.org/10.1016/j.ebiom.2023.104770.
Temel MH, Erden Y, Bağ cıer F. Information quality and readability: ChatGPT’s responses to the most common questions about spinal cord injury. World Neurosurg. 2024;181:e1138–44.
Charnock D, Shepperd S, Needham G, Gann R. DISCERN: an instrument for judging the quality of written consumer health information on treatment choices. J Epidemiol Community Health. 1999;53:105–11. https://doi.org/10.1136/jech.53.2.105.
Funding
The study is supported by the National Natural Science Foundation of China (project code: 81570849 to LPC); Natural Science Foundation of Guangdong Province of China (project code: 2020A1515011413 to LPC); Overseas Famous Teachers Project 2021, Guangdong Province, China (project code:21-294 to LPC).
Author information
Authors and Affiliations
Contributions
ZFF wrote the letter. CLP, HHJ, and LJJ critically revised the letter.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Zhao, FF., He, HJ., Liang, JJ. et al. Reply to ‘Comment on: Benchmarking the performance of large language models in uveitis: a comparative analysis of ChatGPT-3.5, ChatGPT-4.0, Google Gemini, and Anthropic Claude3’. Eye 39, 1433 (2025). https://doi.org/10.1038/s41433-025-03737-x
Received:
Revised:
Accepted:
Published:
Issue date:
DOI: https://doi.org/10.1038/s41433-025-03737-x
This article is cited by
-
Comparison of the readability of ChatGPT and Bard in medical communication: a meta-analysis
BMC Medical Informatics and Decision Making (2025)