论文标题

在内窥镜视觉中进行更好的手术仪器分割:多角度特征聚合和轮廓监督

Towards Better Surgical Instrument Segmentation in Endoscopic Vision: Multi-Angle Feature Aggregation and Contour Supervision

论文作者

Qin, Fangbo, Lin, Shan, Li, Yangming, Bly, Randall A., Moe, Kris S., Hannaford, Blake

论文摘要

准确和实时的手术仪器分割在机器人辅助手术的内窥镜视觉中很重要,并且频繁的仪器组织接触和观察透视的持续改变会带来重大挑战。对于这些具有挑战性的任务,近年来设计了越来越深的神经网络(DNN)模型。我们有动力提出一种可嵌入的方法,以改善这些当前的DNN分割模型,而无需增加模型参数编号。首先,在观察DNN的有限旋转不变性性能时,我们提出了多角度特征聚合(MAFA)方法,利用主动图像旋转来获得更丰富的视觉提示,并使预测更加可靠地对仪器方向变化。其次,在端到端训练阶段,辅助轮廓监督被用来指导模型学习边界意识,以便更精确地分割掩码的轮廓形状。该方法通过从外科医生手术中收集的新型Sinus Surgery数据集上进行了消融实验,并将其与使用DA Vinci XI机器人收集的公共数据集中的现有方法进行了比较。

Accurate and real-time surgical instrument segmentation is important in the endoscopic vision of robot-assisted surgery, and significant challenges are posed by frequent instrument-tissue contacts and continuous change of observation perspective. For these challenging tasks more and more deep neural networks (DNN) models are designed in recent years. We are motivated to propose a general embeddable approach to improve these current DNN segmentation models without increasing the model parameter number. Firstly, observing the limited rotation-invariance performance of DNN, we proposed the Multi-Angle Feature Aggregation (MAFA) method, leveraging active image rotation to gain richer visual cues and make the prediction more robust to instrument orientation changes. Secondly, in the end-to-end training stage, the auxiliary contour supervision is utilized to guide the model to learn the boundary awareness, so that the contour shape of segmentation mask is more precise. The proposed method is validated with ablation experiments on the novel Sinus-Surgery datasets collected from surgeons' operations, and is compared to the existing methods on a public dataset collected with a da Vinci Xi Robot.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源