Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/110106
PIRA download icon_1.1View/Download Full Text
Title: Exploiting cross-scale attention transformer and progressive edge refinement for retinal vessel segmentation
Authors: Yuan, Y
Zhang, Y 
Zhu, L
Cai, L
Qian, Y
Issue Date: Jan-2024
Source: Mathematics, Jan. 2024, v. 12, no. 2, 264
Abstract: Accurate retinal vessel segmentation is a crucial step in the clinical diagnosis and treatment of fundus diseases. Although many efforts have been presented to address the task, the segmentation performance in challenging regions (e.g., collateral vessels) is still not satisfactory, due to their thin morphology or the low contrast between foreground and background. In this work, we observe that an intrinsic appearance exists in the retinal image: among the dendritic vessels there are generous similar structures, e.g., the main and collateral vessels are all curvilinear, but they have noticeable scale differences. Based on this observation, we propose a novel cross-scale attention transformer (CAT) to encourage the segmentation effects in challenging regions. Specifically, CAT consumes features with different scales to produce their shared attention matrix, and then fully integrates the beneficial information between them. Such new attention architecture could explore the multi-scale idea more efficiently, thus realizing mutual learning of similar structures. In addition, a progressive edge refinement module (ERM) is designed to refine the edges of foreground and background in the segmentation results. Through the idea of edge decoupling, ERM could suppress the background feature near the blood vessels while enhancing the foreground feature, so as to segment vessels accurately. We conduct extensive experiments and discussions on DRIVE and CHASE_DB1 datasets to verify the proposed framework. Experimental results show that our method has great advantages in the Se metric, which are 0.88–7.26% and 0.81–7.11% higher than the state-of-the-art methods on DRIVE and CHASE_DB1, respectively. In addition, the proposed method also outperforms other methods with 0.17–2.06% in terms of the Dice metric on DRIVE.
Keywords: Cross-scale attention transformer
Progressive edge refinement
Retinal vessel segmentation
Publisher: MDPI AG
Journal: Mathematics 
EISSN: 2227-7390
DOI: 10.3390/math12020264
Rights: Copyright: © 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
The following publication Yuan Y, Zhang Y, Zhu L, Cai L, Qian Y. Exploiting Cross-Scale Attention Transformer and Progressive Edge Refinement for Retinal Vessel Segmentation. Mathematics. 2024; 12(2):264 is available at https://doi.org/10.3390/math12020264.
Appears in Collections:Journal/Magazine Article

Files in This Item:
File Description SizeFormat 
mathematics-12-00264.pdf3.45 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show full item record

Page views

25
Citations as of Apr 14, 2025

Downloads

11
Citations as of Apr 14, 2025

SCOPUSTM   
Citations

4
Citations as of Sep 12, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.