The automatic segmentation of diabetic retinopathy (DR) holds significant importance for assisting physicians in diagnosis and treatment. Given the complexity, high inter-class similarity, and uncertainty of DR, it is crucial to integrate multiscale information between lesions and establish global correlations among them. To address these issues, a novel HRU-TNet (Hybrid Residual U-Transformer Network) algorithm for retinal lesion segmentation is proposed. In this framework, the network is augmented with lightweight self-attention residual U-modules (LSA-RSU) to capture high-frequency details of the lesions and global contextual information. The skip connections are then enhanced through interactive residual transformer fusion modules (IRTF) and channel-cross attention (CCA), promoting dependencies among features at different scales and filtering out interfering information to guide feature fusion and eliminate ambiguity. Additionally, a novel retinal image enhancement technique is devised, employing local wavelet transformations to capture detailed components of the retinal images, thereby enhancing the representational capacity of the segmentation network. Data augmentation is also performed to ensure network adaptability to small datasets. Comprehensive experiments conducted on the publicly available IDRID and e_ophtha datasets yielded average AUC_PR values of 0.709 and 0.451, respectively. The proposed approach demonstrated superior generalization on the DDR dataset compared to other methods mentioned in the literature. These results demonstrate that our proposed method is better suited for small retinal datasets, exhibiting improved segmentation accuracy and generalization compared to existing approaches.
Read full abstract