Precise Couinaud segmentation from preoperative liver computed tomography (CT) is crucial for surgical planning and lesion examination. However, this task is challenging as it is defined based on vessel structures, and there is no intensity contrast between adjacent Couinaud segments in CT images. To solve this challenge, we design a multi-scale point-voxel fusion framework, which can more effectively model the spatial relationship of points and the semantic information of the image, producing robust and smooth Couinaud segmentations. Specifically, we first segment the liver and vessels from the CT image and generate 3D liver point clouds and voxel grids embedded with the vessel structure. Then, our method with two input-specific branches extracts complementary feature representations from points and voxels, respectively. The local attention module adaptively fuses features from the two branches at different scales to balance the contribution of different branches in learning more discriminative features. Furthermore, we propose a novel distance loss at the feature level to make the features in the segment more compact, thereby improving the certainty of segmentation between segments. Our experimental results on three public liver datasets demonstrate that our proposed method outperforms several state-of-the-art methods by large margins. Specifically, in out-of-distribution (OOD) testing of LiTS dataset, our method exceeded the voxel-based 3D UNet by approximately 20% in Dice score, and outperformed the point-based PointNet2Plus by approximately 8% in Dice score. Our code and manual annotations of the public datasets presented in this paper are available online: https://github.com/xukun-zhang/Couinaud-Segmentation.
Read full abstract