Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/114623
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Electrical and Electronic Engineering-
dc.creatorLai, SJ-
dc.creatorCheung, TH-
dc.creatorFung, KC-
dc.creatorXue, KW-
dc.creatorLam, KM-
dc.date.accessioned2025-08-18T03:02:21Z-
dc.date.available2025-08-18T03:02:21Z-
dc.identifier.issn0277-786X-
dc.identifier.urihttp://hdl.handle.net/10397/114623-
dc.descriptionInternational Workshop on Advanced Imaging Technology (IWAIT) 2025, 6-8 January 2025, Douliu City, Taiwanen_US
dc.language.isoenen_US
dc.publisherSPIE - International Society for Optical Engineeringen_US
dc.rightsCopyright 2024 Society of Photo‑Optical Instrumentation Engineers (SPIE). One print or electronic copy may be made for personal use only. Systematic reproduction and distribution, duplication of any material in this publication for a fee or for commercial purposes, and modification of the contents of the publication are prohibited.en_US
dc.rightsThe following publication Song-Jiang Lai, Tsun-Hin Cheung, Ka-Chun Fung, Kai-wen Xue, and Kin-Man Lam "HAAT: hybrid attention aggregation transformer for image super-resolution", Proc. SPIE 13510, International Workshop on Advanced Imaging Technology (IWAIT) 2025, 135101A (5 February 2025) is available at https://doi.org/10.1117/12.3058003.en_US
dc.subjectAttention mechanismen_US
dc.subjectComputer visionen_US
dc.subjectImage super-resolutionen_US
dc.subjectTransformeren_US
dc.titleHAAT : hybrid attention aggregation transformer for image super-resolutionen_US
dc.typeConference Paperen_US
dc.identifier.volume13510-
dc.identifier.doi10.1117/12.3058003-
dcterms.abstractIn the research area of image super-resolution, Swin-transformer-based models are favored for their global spatial modeling and shifting window attention mechanism. However, existing methods often limit self-attention to nonoverlapping windows to cut costs and ignore the useful information that exists across channels. To address this issue, this paper introduces a novel model, the Hybrid Attention Aggregation Transformer (HAAT), designed to better leverage feature information. HAAT is constructed by integrating Swin-Dense-Residual-Connected Blocks (SDRCB) with Hybrid Grid Attention Blocks (HGAB). SDRCB expands the receptive field while maintaining a streamlined architecture, resulting in enhanced performance. HGAB incorporates channel attention, sparse attention, and window attention to improve nonlocal feature fusion and achieve more visually compelling results. Experimental evaluations demonstrate that HAAT surpasses state-of-the-art methods on benchmark datasets.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationProceedings of SPIE : the International Society for Optical Engineering, 2025, v. 13510, 135101A-
dcterms.isPartOfProceedings of SPIE : the International Society for Optical Engineering-
dcterms.issued2025-
dc.identifier.scopus2-s2.0-85218338826-
dc.relation.conferenceInternational Workshop on Advanced Imaging Technology [IWAIT]-
dc.identifier.eissn1996-756X-
dc.identifier.artn135101A-
dc.description.validate202508 bcch-
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumberOA_Othersen_US
dc.description.fundingSourceSelf-fundeden_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryVoR alloweden_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
135101A.pdf911.56 kBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.