2024 Journal of oral and maxillofa…

Is ChatGPT an Accurate and Readable Patient Aid for Third Molar Extractions?

, , ,

Journal of oral and maxillofacial surgery : official journal of the American Association of Oral and Maxillofacial Surgeons Vol. 82 (10) : 1239-1245 • Oct 2024

BACKGROUND: Artificial intelligence (AI) platforms such as Chat Generative Pre-Trained Transformer (ChatGPT) (Open AI, San Francisco, California, USA) have the capacity to answer health-related questions. It remains unknown whether AI can be a patient-friendly and accurate resource regarding third molar extraction. PURPOSE: The purpose was to determine the accuracy and readability of AI responses to common patient questions regarding third molar extraction. STUDY DESIGN, SETTING, SAMPLE: This is a cross sectional in-silico assessment of readability and soundness of a computer-generated report. INDEPENDENT VARIABLE: Not applicable. MAIN OUTCOME VARIABLES: Accuracy, or the ability to provide clinically correct and relevant information, was determined subjectively by 2 reviewers using a 5-point Likert scale, and objectively by comparing responses to American Association of Oral and Maxillofacial Surgeons (AAOMS) clinical consensus papers. Readability, or how easy a piece of text is to read, was assessed using the Flesch Kincaid Reading Ease (FKRE) and Flesch Kincaid Grade Level (FKGL). Both assess readability based on mean number of syllables per word, and words per sentence. To be deemed readable, FKRE should be >60 and FKGL should be <8. COVARIATES: Not applicable. ANALYSES: Descriptive statistics were used to analyze the findings of this study. RESULTS: AI-generated responses above the recommended level for the average patient (FKRE: 52; FKGL: 10). The average Likert score was 4.36, suggesting that most responses were accurate with minor inaccuracies or missing information. AI correctly deferred to the provider in instances where no definitive answer exists. Of the responses that addressed content in AAOMS consensus papers, 18/19 responses closely aligned with them. All prompts did not provide citations or references. CONCLUSION AND RELEVANCE: AI was able to provide mostly accurate responses, and content was closely aligned with AAOMS guidelines. However, responses were too complex for the average third molar extraction patient, and were deficient in citations and references. It is important for providers to educate patients on the utility of AI, and to decide whether to recommend using it for information. Ultimately, the best resource for answers is from the practitioners themselves because the AI platform lacks clinical experience.

No clinical trial protocols linked to this paper

Clinical trials are automatically linked when NCT numbers are found in the paper's title or abstract.
PICO Elements

No PICO elements extracted yet. Click "Extract PICO" to analyze this paper.

Paper Details
MeSH Terms
Associated Data

No associated datasets or code repositories found for this paper.

Related Papers

Related paper suggestions will be available in future updates.