Articulated Multi-Instrument 2-D Pose Estimation Using Fully Convolutional Networks.

Research paper by Xiaofei X Du, Thomas T Kurmann, Ping-Lin PL Chang, Maximilian M Allan, Sebastien S Ourselin, Raphael R Sznitman, John D JD Kelly, Danail D Stoyanov

Indexed on: 05 May '18Published on: 05 May '18Published in: IEEE transactions on medical imaging


Instrument detection, pose estimation, and tracking in surgical videos are an important vision component for computer-assisted interventions. While significant advances have been made in recent years, articulation detection is still a major challenge. In this paper, we propose a deep neural network for articulated multi-instrument 2-D pose estimation, which is trained on detailed annotations of endoscopic and microscopic data sets. Our model is formed by a fully convolutional detection-regression network. Joints and associations between joint pairs in our instrument model are located by the detection subnetwork and are subsequently refined through a regression subnetwork. Based on the output from the model, the poses of the instruments are inferred using maximum bipartite graph matching. Our estimation framework is powered by deep learning techniques without any direct kinematic information from a robot. Our framework is tested on single-instrument RMIT data, and also on multi-instrument EndoVis and in vivo data with promising results. In addition, the data set annotations are publicly released along with our code and model.