Abstract

Surgical instrument segmentation plays a crucial role in robot-assisted surgery by furnishing essential information about instrument location and orientation. This information not only enhances surgical planning but also augments the precision and safety of procedures. Despite promising strides in recent research on surgical instrument segmentation, accuracy still faces obstacles due to local feature processing limitations, surgical environment complexity, and instrument morphological variability. To address these challenges, we introduced the channel-wise features fusion and recalibration network (CFFR-Net). This network utilizes a dual-stream mechanism, combining a context-guided block and dense block for feature extraction. The context-guided block captures a variety of contextual information by using different dilation rates. Additionally, CFFR-Net employs a fusion mechanism that harmonizes context-guided and dense streams. This integration, along with the inclusion of Squeeze-and-Excitation attention, enhances both the precision and robustness of semantic instrument segmentation.We performed experiments using two publicly available datasets for surgical instrument segmentation: the Kvasir-instrument and Endovis2017 datasets. The results of these experiments were highly encouraging, as our proposed model exhibited remarkable performance on both datasets compared to the state-of-the-art methods. On the Kvasir-instrument set, our model achieved a Dice score of 95.84% and mean intersection over union (mIOU) value of 92.40%. Similarly, on the Endovis2017 set, it obtained a Dice score of 95.47% and mIOU value of 93.02%.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call