SuFIA: Language-Guided Augmented Dexterity
for Robotic Surgical Assistants

 
1University of Toronto, 2University of Bern, 3University of California, Berkeley,
4NVIDIA, 5Georgia Institute of Technology
 
IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2024

Abstract

In this work, we present SuFIA, the first framework for natural language-guided augmented dexterity for robotic surgical assistants. SuFIA incorporates the strong reasoning capabilities of large language models (LLMs) with perception modules to implement high-level planning and low-level control of a robot for surgical sub-task execution. This enables a learning-free approach to surgical augmented dexterity without any in-context examples or motion primitives. SuFIA uses a human-in-the-loop paradigm by restoring control to the surgeon in the case of insufficient information, mitigating unexpected errors for mission-critical tasks. We evaluate SuFIA on four surgical sub-tasks in a simulation environment and two sub-tasks on a physical surgical robotic platform in the lab, demonstrating its ability to perform common surgical sub-tasks through supervised autonomous operation under challenging physical and workspace conditions.

Video

Example Tasks

Suture Needle Handover

In this task, the robot hands over a suture needle from one arm to the other. SuFIA queries a language model to understand the surgeon's request and plan the handover motion and also communicates its intentions to the surgeon. SuFIA directly devises and executes the low-level robot actions and trajectories for the handover motion. The robot can adapt to the surgeon's preferences and adjust the handover motion accordingly.

Vessel Dilation

In this task, a spring clamp assembly holds a soft vessel phantom from two points. The dVRK arm is required to grip the vessel rim from a third point facing the robot and dilate the vessel by pulling backward.

Our Related Projects

SuFIA-BC: Generating High Quality Demonstration Data for Visuomotor Policy Learning in Surgical Subtasks
Masoud Moghani, Nigel Nelson, Mohamed Ghanem, Andres Diaz-Pinto,
Kush Hari, Mahdi Azizian, Ken Goldberg, Sean Huver, Animesh Garg
website / video

We present SuFIA-BC: exploring visual Behavior Cloning policies for Surgical First Interactive Autonomy Assistants. We provide an enhanced surgical digital twin with photorealistic human anatomical organs, integrated into ORBIT-Surgical designed to generate high-quality synthetic data for solving fundamental tasks in surgical autonomy. We investigate visual observation spaces including multi-view cameras and 3D visual representations extracted from a single endoscopic camera view.