1. An investigative analysis - ChatGPT's capability to excel in the Polish speciality exam in pathology.
- Author
-
Bielówka M, Kufel J, Rojek M, Kaczyńska D, Czogalik Ł, Mitręga A, Bartnikowska W, Kondoł D, Palkij K, and Mielcarska S
- Subjects
- Humans, Poland, Educational Measurement methods, Artificial Intelligence, Pathology education
- Abstract
This study evaluates the effectiveness of the ChatGPT-3.5 language model in providing correct answers to pathomorphology questions as required by the State Speciality Examination (PES). Artificial intelligence (AI) in medicine is generating increasing interest, but its potential needs thorough evaluation. A set of 119 exam questions by type and subtype were used, which were posed to the ChatGPT-3.5 model. Performance was analysed with regard to the success rate in different question categories and subtypes. ChatGPT-3.5 achieved a performance of 45.38%, which is significantly below the minimum PES pass threshold. The results achieved varied by question type and subtype, with better results in questions requiring "comprehension and critical thinking" than "memory". The analysis shows that, although ChatGPT-3.5 can be a useful teaching tool, its performance in providing correct answers to pathomorphology questions is significantly lower than that of human respondents. This conclusion highlights the need to further improve the AI model, taking into account the specificities of the medical field. Artificial intelligence can be helpful, but it cannot fully replace the experience and knowledge of specialists.
- Published
- 2024
- Full Text
- View/download PDF