Comparative analysis of artificial intelligence chatbots in orthodontic emergency scenarios: ChatGPT-3.5, ChatGPT-4.0, Copilot, and Gemini
Tarih
Yazarlar
Dergi Başlığı
Dergi ISSN
Cilt Başlığı
Yayıncı
Erişim Hakkı
Özet
Objectives To evaluate and compare the accuracy of four AI chatbots, ChatGPT-3.5, ChatGPT-4.0, Copilot, and Gemini, in response to orthodontic emergency scenarios. Materials and Methods Forty frequently asked questions related to orthodontic emergencies were posed to the chatbots. These questions were categorized as fixed orthodontic treatment, clear aligner treatment, eating and oral hygiene, pain and discomfort, general concerns, retention, and sports and travel. The responses were evaluated by three orthodontic experts using a five-point Likert scale, and statistical analysis was conducted to assess variations in accuracy across chatbots. Results Statistical analysis revealed significant differences among the chatbots. Gemini and ChatGPT-4.0 demonstrated the highest accuracy in response to orthodontic emergencies, followed by Copilot, whereas ChatGPT-3.5 had the lowest accuracy scores. Additionally, the “Fixed Orthodontic Treatment” category showed a statistically significant difference (P = .043), with Gemini outperforming the other chatbots in this category. However, no statistically significant differences were found in other categories. Conclusions AI chatbots show potential in providing immediate assistance for orthodontic emergencies, but their accuracy varies across different models and question categories.