Can artificial intelligence pass the written European Board of Hand Surgery exam? - 13/09/25
, Lindsay Muir b, c, Benedikt Fuchs a, Vanessa Lucksch a, Felix H. Vollbach a, Elisabeth M. Haas-Lützenberger a, Riccardo E. Giunta a, Nikolaus Thierfelder a, Wolfram Demmer aAbstract |
Various artificial intelligence-based applications have emerged as transformative tools across numerous domains. Among these, ChatGPT has earned global recognition with its capacity for dynamic user interaction and holds significant potential in the medical sector. However, the subject-specific accuracy of ChatGPT remains a matter of debate.
This study assesses the capabilities and knowledge of different artificial intelligence chatbots (ChatGPT, Google Gemini, and Claude) in the domain of hand surgery. Each chatbot conducted a full written EBHS exam. The test results were analyzed according to the EBHS-guidelines, focused on the total scores and the ratio of correct to incorrect responses for each artificial intelligence model. Findings revealed that three out of the four chatbots achieved passing scores on the exam. Notably, ChatGPT-4o1 demonstrated significantly superior performance.
This study highlights the subject-specific expertise of different artificial intelligence programs within the specialized field of hand surgery while also underscoring their variability and limitations.
Le texte complet de cet article est disponible en PDF.Keywords : Artificial intelligence, ChatGPT, Google Gemini, Claude, EBHS exam, FESSH
Plan
Vol 44 - N° 4
Article 102197- septembre 2025 Retour au numéroBienvenue sur EM-consulte, la référence des professionnels de santé.
L’accès au texte intégral de cet article nécessite un abonnement.
Déjà abonné à cette revue ?
