1. ChatGPT compared to national guidelines for management of ovarian cancer: Did ChatGPT get it right? – A Memorial Sloan Kettering Cancer Center Team Ovary study.
- Author
-
Finch, Lindsey, Broach, Vance, Feinberg, Jacqueline, Al-Niaimi, Ahmed, Abu-Rustum, Nadeem R., Zhou, Qin, Iasonos, Alexia, and Chi, Dennis S.
- Subjects
- *
LANGUAGE models , *CHATGPT , *CHATBOTS , *ARTIFICIAL intelligence , *GENERATIVE pre-trained transformers - Abstract
We evaluated the performance of a chatbot compared to the National Comprehensive Cancer Network (NCCN) Guidelines for the management of ovarian cancer. Using NCCN Guidelines, we generated 10 questions and answers regarding management of ovarian cancer at a single point in time. Questions were thematically divided into risk factors, surgical management, medical management, and surveillance. We asked ChatGPT (GPT-4) to provide responses without prompting (unprompted GPT) and with prompt engineering (prompted GPT). Responses were blinded and evaluated for accuracy and completeness by 5 gynecologic oncologists. A score of 0 was defined as inaccurate, 1 as accurate and incomplete, and 2 as accurate and complete. Evaluations were compared among NCCN, unprompted GPT, and prompted GPT answers. Overall, 48% of responses from NCCN, 64% from unprompted GPT, and 66% from prompted GPT were accurate and complete. The percentage of accurate but incomplete responses was higher for NCCN vs GPT-4. The percentage of accurate and complete scores for questions regarding risk factors, surgical management, and surveillance was higher for GPT-4 vs NCCN; however, for questions regarding medical management, the percentage was lower for GPT-4 vs NCCN. Overall, 14% of responses from unprompted GPT, 12% from prompted GPT, and 10% from NCCN were inaccurate. GPT-4 provided accurate and complete responses at a single point in time to a limited set of questions regarding ovarian cancer, with best performance in areas of risk factors, surgical management, and surveillance. Occasional inaccuracies, however, should limit unsupervised use of chatbots at this time. • There is interest in expanding the use of chatbot technology in medicine. • A high percentage of responses from ChatGPT to questions about ovarian cancer were graded as accurate and complete. • ChatGPT had a higher percentage of inaccurate responses compared to the National Comprehensive Cancer Network Guidelines. • ChatGPT received the highest scores in response to questions regarding risk factors, surgical management, and surveillance. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF