Abstract

One-to-multiple medical image segmentation aims to directly test a segmentation model trained with the medical images of a one-domain site on those of a multiple-domain site, suffering from segmentation performance degradation on multiple domains. This process avoids additional annotations and helps improve the application value of the model. However, no successful one-to-multiple unsupervised domain adaptation (O2M-UDA) work has been reported in one-to-multiple medical image segmentation due to its inherent challenges: distribution differences among multiple target domains (among-target differences) caused by different scanning equipment and distribution differences between one source domain and multiple target domains (source–target differences). In this paper, we propose an O2M-UDA framework called dynamic domain adaptation (DyDA), for one-to-multiple medical image segmentation, which has two innovations: (1) dynamic credible sample strategy (DCSS) dynamically extracts credible samples from the target site and iteratively updates their number, thus iteratively expanding the generalization boundary of the model and minimizing the among-target differences; (2) hybrid uncertainty learning (HUL) reduces the voxel-level and domain-level uncertainty simultaneously, thus minimizing the source–target differences from the detail and entire perspective concurrently. Experiments on two one-to-multiple medical image segmentation tasks have been conducted to demonstrate the performance of the proposed DyDA. The proposed DyDA achieved competitive segmentation results and high adaptation with an average of 83.8% and 48.1% dice for the two tasks, respectively, which has improved by 21.7% and 9.2% compared with no adaptation, respectively. The code developed in this study code can be downloaded at https://github.com/ZoeyJiang/DyDA.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call