1. A Dataset and Benchmarks for Segmentation and Recognition of Gestures in Robotic Surgery
- Author
-
Narges Ahmidi, Lingling Tao, Shahin Sefati, Sanjeev Khudanpur, Colin Lea, Gregory D. Hager, Yixin Gao, Luca Zappella, René Vidal, and Benjamin Bejar Haro
- Subjects
Conditional random field ,Engineering ,Databases, Factual ,Biomedical Engineering ,02 engineering and technology ,Machine learning ,computer.software_genre ,Article ,Field (computer science) ,Pattern Recognition, Automated ,030218 nuclear medicine & medical imaging ,Activity recognition ,03 medical and health sciences ,Imaging, Three-Dimensional ,0302 clinical medicine ,Robotic Surgical Procedures ,0202 electrical engineering, electronic engineering, information engineering ,Feature (machine learning) ,Humans ,Segmentation ,Hidden Markov model ,Gestures ,Markov chain ,business.industry ,United States ,Benchmarking ,020201 artificial intelligence & image processing ,Clinical Competence ,Artificial intelligence ,business ,computer ,Gesture - Abstract
Objective : State-of-the-art techniques for surgical data analysis report promising results for automated skill assessment and action recognition. The contributions of many of these techniques, however, are limited to study-specific data and validation metrics, making assessment of progress across the field extremely challenging. Methods : In this paper, we address two major problems for surgical data analysis: First, lack of uniform-shared datasets and benchmarks, and second, lack of consistent validation processes. We address the former by presenting the JHU-ISI Gesture and Skill Assessment Working Set (JIGSAWS), a public dataset that we have created to support comparative research benchmarking. JIGSAWS contains synchronized video and kinematic data from multiple performances of robotic surgical tasks by operators of varying skill. We address the latter by presenting a well-documented evaluation methodology and reporting results for six techniques for automated segmentation and classification of time-series data on JIGSAWS. These techniques comprise four temporal approaches for joint segmentation and classification: hidden Markov model, sparse hidden Markov model (HMM), Markov semi-Markov conditional random field, and skip-chain conditional random field; and two feature-based ones that aim to classify fixed segments: bag of spatiotemporal features and linear dynamical systems. Results : Most methods recognize gesture activities with approximately 80% overall accuracy under both leave-one-super-trial-out and leave-one-user-out cross-validation settings. Conclusion : Current methods show promising results on this shared dataset, but room for significant progress remains, particularly for consistent prediction of gesture activities across different surgeons. Significance : The results reported in this paper provide the first systematic and uniform evaluation of surgical activity recognition techniques on the benchmark database.
- Published
- 2017
- Full Text
- View/download PDF