PaI‐Net: A modified U‐Net of reducing semantic gap for surgical instrument segmentation

Tracking the instruments in a surgical scene is an essential task in minimally invasive surgery. However, due to the unpredictability of scenes, automatically segmenting the instruments is very challenging. In this paper, a novel method named parallel inception network (PaI‐Net) is proposed, in whic...

Full description

Saved in:
Bibliographic Details
Published inIET image processing Vol. 15; no. 12; pp. 2959 - 2969
Main Authors Wang, Xiaoyan, Wang, Luyao, Zhong, Xingyu, Bai, Cong, Huang, Xiaojie, Zhao, Ruiyi, Xia, Ming
Format Journal Article
LanguageEnglish
Published Wiley 01.10.2021
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Tracking the instruments in a surgical scene is an essential task in minimally invasive surgery. However, due to the unpredictability of scenes, automatically segmenting the instruments is very challenging. In this paper, a novel method named parallel inception network (PaI‐Net) is proposed, in which an attention parallel module (APM) and an output fusion module (OFM) are integrated with U‐Net to improve the segmentation ability. Specially, APM utilizes multi‐scale convolution kernels and global average pooling operations to extract semantic information and global context information of different scales, while OFM combines the feature maps of the decoder part to aggregate the abundant boundary information of shallow layers and the rich semantic information of deep layers together, which achieve a significant improvement in generating segmentation masks. Finally, the evaluation of proposed method on robotic instruments segmentation task from Medical Image Computing and Computer Assisted Intervention Society (MICCAI) and retinal image segmentation task from International Symposium on Biomedical Imaging (ISBI) show that our model has achieved advanced performance on multi‐scale semantic segmentation and is superior to the current state‐of‐the‐art models.
ISSN:1751-9659
1751-9667
DOI:10.1049/ipr2.12283