daVinciNet: Joint Prediction of Motion and Surgical State in Robot-Assisted Surgery
Abstract
This paper presents a technique to concurrently and jointly predict the future trajectories of surgical instruments and the future state(s) of surgical subtasks in robot-assisted surgeries (RAS) using multiple input sources. Such predictions are a necessary first step towards shared control and supervised autonomy of surgical subtasks. Minute-long surgical subtasks, such as suturing or ultrasound scanning, often have distinguishable tool kinematics and visual features, and can be described as a series of fine-grained states with transition schematics. We propose daVinciNet - an end-to-end dual-task model for robot motion and surgical state predictions. daVinciNet performs concurrent end-effector trajectory and surgical state predictions using features extracted from multiple data streams, including robot kinematics, endoscopic vision, and system events. We evaluate our proposed model on an extended Robotic Intra-Operative Ultrasound (RIOUS+) imaging dataset collected on a da Vinci® Xi surgical system and the JHU-ISI Gesture and Skill Assessment Working Set (JIGSAWS). Our model achieves up to 93.85% short-term (0.5s) and 82.11% long-term (2s) state prediction accuracy, as well as 1.07mm short-term and 5.62mm long-term trajectory prediction error.
Additional Information
© 2020 IEEE. This work was funded by Intuitive Surgical, Inc. We would like to thank Dr. Azad Shademan and Dr. A. Jonathan McLeod for their support of this research.Attached Files
Accepted Version - 2009.11937.pdf
Files
Name | Size | Download all |
---|---|---|
md5:23655d1c7f9c77fe26e0cfb428be159e
|
7.1 MB | Preview Download |
Additional details
- Eprint ID
- 107578
- Resolver ID
- CaltechAUTHORS:20210119-161653290
- Intuitive Surgical, Inc.
- Created
-
2021-01-20Created from EPrint's datestamp field
- Updated
-
2021-11-16Created from EPrint's last_modified field