Skip to main content
Top

20-05-2024 | Osteoporosis | Original Article

Educating patients on osteoporosis and bone health: Can “ChatGPT” provide high-quality content?

Authors: Diane Ghanem, Henry Shu, Victoria Bergstein, Majd Marrache, Andra Love, Alice Hughes, Rachel Sotsky, Babar Shafiq

Published in: European Journal of Orthopaedic Surgery & Traumatology

Login to get access

Abstract

Purpose

The rise of artificial intelligence (AI) models like ChatGPT offers potential for varied applications, including patient education in healthcare. With gaps in osteoporosis and bone health knowledge and adherence to prevention and treatment, this study aims to evaluate the accuracy of ChatGPT in delivering evidence-based information related to osteoporosis.

Methods

Twenty of the most common frequently asked questions (FAQs) related to osteoporosis were subcategorized into diagnosis, diagnostic method, risk factors, and treatment and prevention. These FAQs were sourced online and inputted into ChatGPT-3.5. Three orthopedic surgeons and one advanced practice provider who routinely treat patients with fragility fractures independently reviewed the ChatGPT-generated answers, grading them on a scale from 0 (harmful) to 4 (excellent). Mean response accuracy scores were calculated. To compare the variance of the means across the four categories, a one-way analysis of variance (ANOVA) was used.

Results

ChatGPT displayed an overall mean accuracy score of 91%. Its responses were graded as “accurate requiring minimal clarification” or “excellent,” with a mean response score ranging from 3.25 to 4. No answers were deemed inaccurate or harmful. No significant difference was observed in the means of responses across the defined categories.

Conclusion

ChatGPT-3.5 provided high-quality educational content. It showcased a high degree of accuracy in addressing osteoporosis-related questions, aligning closely with expert opinions and current literature, with structured and inclusive answers. However, while AI models can enhance patient information accessibility, they should be used as an adjunct rather than a substitute for human expertise and clinical judgment.
Appendix
Available only for authorised users
Literature
7.
go back to reference Kassab J, El Dahdah J, Chedid El Helou M, Layoun H, Sarraju A, Laffin LJ, Harb SC (2023) Assessing the accuracy of an online chat-based artificial intelligence model in providing recommendations on hypertension management in accordance with the 2017 american college of cardiology/american heart association and 2018 european society of cardiology/european society of hypertension guidelines. Hypertension 80(7):e125–e127CrossRefPubMed Kassab J, El Dahdah J, Chedid El Helou M, Layoun H, Sarraju A, Laffin LJ, Harb SC (2023) Assessing the accuracy of an online chat-based artificial intelligence model in providing recommendations on hypertension management in accordance with the 2017 american college of cardiology/american heart association and 2018 european society of cardiology/european society of hypertension guidelines. Hypertension 80(7):e125–e127CrossRefPubMed
14.
go back to reference Open AI (2023) GPT-4 Technical Report. Published online 27 March 2023 Open AI (2023) GPT-4 Technical Report. Published online 27 March 2023
Metadata
Title
Educating patients on osteoporosis and bone health: Can “ChatGPT” provide high-quality content?
Authors
Diane Ghanem
Henry Shu
Victoria Bergstein
Majd Marrache
Andra Love
Alice Hughes
Rachel Sotsky
Babar Shafiq
Publication date
20-05-2024
Publisher
Springer Paris
Published in
European Journal of Orthopaedic Surgery & Traumatology
Print ISSN: 1633-8065
Electronic ISSN: 1432-1068
DOI
https://doi.org/10.1007/s00590-024-03990-y