Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/106880
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Electrical and Electronic Engineeringen_US
dc.creatorLiu, Ten_US
dc.creatorZhao, Ren_US
dc.creatorLam, KMen_US
dc.date.accessioned2024-06-07T00:58:36Z-
dc.date.available2024-06-07T00:58:36Z-
dc.identifier.isbn978-1-5106-4364-2en_US
dc.identifier.isbn978-1-5106-4365-9 (electronic)en_US
dc.identifier.issn0277-786Xen_US
dc.identifier.urihttp://hdl.handle.net/10397/106880-
dc.descriptionInternational Workshop on Advanced Imaging Technology 2021 (IWAIT 2021), 2021, Online Onlyen_US
dc.language.isoenen_US
dc.publisherSPIE - International Society for Optical Engineeringen_US
dc.rights© (2021) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). One print or electronic copy may be made for personal use only. Systematic reproduction and distribution, duplication of any material in this publication for a fee or for commercial purposes, and modification of the contents of the publication are prohibited.en_US
dc.rightsThe following publication Tianshan Liu, Rui Zhao, and Kin-Man Lam "Attention-based cross-modality interaction for multispectral pedestrian detection", Proc. SPIE 11766, International Workshop on Advanced Imaging Technology (IWAIT) 2021, 1176603 (13 March 2021) is available at https://doi.org/10.1117/12.2590661.en_US
dc.titleAttention-based cross-modality interaction for multispectral pedestrian detectionen_US
dc.typeConference Paperen_US
dc.identifier.volume11766en_US
dc.identifier.doi10.1117/12.2590661en_US
dcterms.abstractMultispectral pedestrian detection has attracted extensive attention, as paired RGB-thermal images can provide complementary patterns to deal with illumination changes in realistic scenarios. However, most of the existing deep-learning-based multispectral detectors extract features from RGB and thermal inputs separately, and fuse them by a simple concatenation operation. This fusion strategy is suboptimal, as undifferentiated concatenation for each region and feature channel may hamper the optimal selection of complementary features from different modalities. To address this limitation, in this paper, we propose an attention-based cross-modality interaction (ACI) module, which aims to adaptively highlight and aggregate the discriminative regions and channels of the feature maps from RGB and thermal images. The proposed ACI module is deployed into multiple layers of a two-branch-based deep architecture, to capture the cross-modal interactions from diverse semantic levels, for illumination-invariant pedestrian detection. Experimental results on the public KAIST multispectral pedestrian benchmark show that the proposed method achieves state-of-the-art detection performance.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationProceedings of SPIE : the International Society for Optical Engineering, 2021, v. 11766, 1176603en_US
dcterms.isPartOfProceedings of SPIE : the International Society for Optical Engineeringen_US
dcterms.issued2021-
dc.identifier.scopus2-s2.0-85103288288-
dc.relation.conferenceInternational Workshop on Advanced Imaging Technology [IWAIT]en_US
dc.identifier.eissn1996-756Xen_US
dc.identifier.artn1176603en_US
dc.description.validate202405 bcchen_US
dc.description.oaAccepted Manuscripten_US
dc.identifier.FolderNumberEIE-0076-
dc.description.fundingSourceSelf-fundeden_US
dc.description.pubStatusPublisheden_US
dc.identifier.OPUS53438687-
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
Liu_Attention-Based_Cross-Modality_Interaction.pdfPre-Published version1.01 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

81
Last Week
3
Last month
Citations as of Nov 9, 2025

Downloads

60
Citations as of Nov 9, 2025

SCOPUSTM   
Citations

3
Citations as of Dec 19, 2025

WEB OF SCIENCETM
Citations

3
Citations as of Dec 18, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.