School of Medicine Publications and Presentations
Document Type
Article
Publication Date
10-2025
Abstract
Purpose: The study aimed to evaluate the accuracy, comprehensiveness, and readability of responses generated by ChatGPT 4.0 to 30 common patient questions about the Bernese periacetabular osteotomy (PAO).
Methods: Two fellowship-trained orthopaedic surgeons specializing in hip preservation selected thirty questions from a prior study identifying common PAO questions on social media. Each question was entered into ChatGPT 4.0, and the surgeons independently graded responses. Responses were evaluated using an established grading system. Accuracy and comprehensiveness were assessed based on the concordance of response content with current literature. Readability was analysed by calculating the Flesch-Kincaid Grade Level and Flesch-Kincaid Reading Ease.
Results: Regarding accuracy and comprehensiveness, 98.3% of responses were graded as "excellent" or "satisfactory, requiring minimal clarification." Readability analysis revealed an average Flesch-Kincaid Grade Level corresponding to an 11th-grade reading level (11.09 ± 1.47) and a mean Reading Ease score requiring college level reading comprehension (39.12 ± 8.25) for original responses, 8th-grade reading level (8.16 ± 1.46) requiring high school to college level reading comprehension (51.53 ± 9.62) for simplified responses, and 7th-grade reading level (7.09 ± 1.23) requiring high school level reading comprehension (62.46 ± 7.48) for 6th grade responses.
Conclusion: ChatGPT 4.0 offered excellent or satisfactory answers to the most common questions surrounding PAO. Asking ChatGPT 4.0 to simplify or respond at a specific reading level may increase the readability of responses. The 4.0 model has shown the potential to be a valuable adjunct for patient education, though the readability may need to be improved via simplified responses.
Level of evidence: Level N/A.
Recommended Citation
Gaddis, J. M., Arellano, E., Martin, B. C., Alsabawi, Y., Salgado‐Flores, M., South, C., ... & Wells, J. E. (2025). Assessing the accuracy and readability of ChatGPT 4.0's original and simplified responses to common patient questions regarding periacetabular osteotomy. Journal of Experimental Orthopaedics, 12(4), e70457. https://doi.org/10.1002/jeo2.70457
Creative Commons License

This work is licensed under a Creative Commons Attribution 4.0 International License.
DOI
Journal of Experimental Orthopaedics
Academic Level
medical student

Comments
© 2025 The Author(s). Journal of Experimental Orthopaedics published by John Wiley & Sons Ltd on behalf of European Society of Sports Traumatology, Knee Surgery and Arthroscopy. This is an open access article under the terms of the http://creativecommons.org/licenses/by/4.0/ License, which permits use, distribution and reproduction in any medium, provided the original work is properly cited.