Welcome to the new version of CaltechAUTHORS. Login is currently restricted to library staff. If you notice any issues, please email coda@library.caltech.edu
Published October 2020 | Accepted Version
Book Section - Chapter Open

daVinciNet: Joint Prediction of Motion and Surgical State in Robot-Assisted Surgery

Abstract

This paper presents a technique to concurrently and jointly predict the future trajectories of surgical instruments and the future state(s) of surgical subtasks in robot-assisted surgeries (RAS) using multiple input sources. Such predictions are a necessary first step towards shared control and supervised autonomy of surgical subtasks. Minute-long surgical subtasks, such as suturing or ultrasound scanning, often have distinguishable tool kinematics and visual features, and can be described as a series of fine-grained states with transition schematics. We propose daVinciNet - an end-to-end dual-task model for robot motion and surgical state predictions. daVinciNet performs concurrent end-effector trajectory and surgical state predictions using features extracted from multiple data streams, including robot kinematics, endoscopic vision, and system events. We evaluate our proposed model on an extended Robotic Intra-Operative Ultrasound (RIOUS+) imaging dataset collected on a da Vinci® Xi surgical system and the JHU-ISI Gesture and Skill Assessment Working Set (JIGSAWS). Our model achieves up to 93.85% short-term (0.5s) and 82.11% long-term (2s) state prediction accuracy, as well as 1.07mm short-term and 5.62mm long-term trajectory prediction error.

Additional Information

© 2020 IEEE. This work was funded by Intuitive Surgical, Inc. We would like to thank Dr. Azad Shademan and Dr. A. Jonathan McLeod for their support of this research.

Attached Files

Accepted Version - 2009.11937.pdf

Files

2009.11937.pdf
Files (7.1 MB)
Name Size Download all
md5:23655d1c7f9c77fe26e0cfb428be159e
7.1 MB Preview Download

Additional details

Created:
August 19, 2023
Modified:
October 23, 2023