Abstract
Retinal fundus images serve as a non-invasive modality to obtain information pertaining to retinal vessels through fundus photography, thereby offering insights into cardiovascular and cerebrovascular diseases. Retinal arteriolar morphometry has emerged as the most convenient and fundamental clinical methodology in the realm of patient screening and diagnosis. Nevertheless, the analysis of retinal arterioles is challenging attributable to imaging noise, stochastic fuzzy characteristics, and blurred boundaries proximal to blood vessels. In response to these limitations, we introduce an innovative methodology, named PKSEA-Net, which aims to improve segmentation accuracy by enhancing the perception of edge information in retinal fundus images. PKSEA-Net employs the universal architecture PVT-v2 as the encoder, complemented by a novel decoder architecture consisting of an Edge-Aware Block (EAB) and a Pyramid Feature Fusion Module (PFFM). The EAB block incorporates prior knowledge for supervision and multi-query for multi-task learning, with supervision information derived from an enhanced Full Width at Half Maximum (FWHM) algorithm and gradient map. Moreover, PFFM efficiently integrates multi-scale features through a novel attention fusion method. Additionally, we have collected a Retinal Cross-Sectional Vessel (RCSV) dataset derived from approximately 200 patients in Quzhou People’s Hospital to serve as the benchmark dataset. Comparative evaluations with several state-of-the-art (SOTA) networks confirm that PKSEA-Net achieves exceptional experimental performance, thereby establishing its status as a SOTA approach for precise boundary delineation and retinal vessel segmentation.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.