Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/108819
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Electrical and Electronic Engineeringen_US
dc.creatorCai, Cen_US
dc.creatorWang, Yen_US
dc.creatorYap, KHen_US
dc.date.accessioned2024-08-27T04:40:47Z-
dc.date.available2024-08-27T04:40:47Z-
dc.identifier.urihttp://hdl.handle.net/10397/108819-
dc.language.isoenen_US
dc.publisherMDPI AGen_US
dc.rights© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).en_US
dc.rightsThe following publication Cai C, Wang Y, Yap K-H. Interactive Change-Aware Transformer Network for Remote Sensing Image Change Captioning. Remote Sensing. 2023; 15(23):5611 is available at https://doi.org/10.3390/rs15235611.en_US
dc.subjectImage change captioningen_US
dc.subjectMulti-layer change awarenessen_US
dc.subjectRemote sensingen_US
dc.subjectTransformeren_US
dc.titleInteractive change-aware transformer network for remote sensing image change captioningen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume15en_US
dc.identifier.issue23en_US
dc.identifier.doi10.3390/rs15235611en_US
dcterms.abstractRemote sensing image change captioning (RSICC) aims to automatically generate sentences describing the difference in content in remote sensing bitemporal images. Recent works extract the changes between bitemporal features and employ a hierarchical approach to fuse multiple changes of interest, yielding change captions. However, these methods directly aggregate all features, potentially incorporating non-change-focused information from each encoder layer into the change caption decoder, adversely affecting the performance of change captioning. To address this problem, we proposed an Interactive Change-Aware Transformer Network (ICT-Net). ICT-Net is able to extract and incorporate the most critical changes of interest in each encoder layer to improve change description generation. It initially extracts bitemporal visual features from the CNN backbone and employs an Interactive Change-Aware Encoder (ICE) to capture the crucial difference between these features. Specifically, the ICE captures the most change-aware discriminative information between the paired bitemporal features interactively through difference and content attention encoding. A Multi-Layer Adaptive Fusion (MAF) module is proposed to adaptively aggregate the relevant change-aware features in the ICE layers while minimizing the impact of irrelevant visual features. Moreover, we extend the ICE to extract multi-scale changes and introduce a novel Cross Gated-Attention (CGA) module into the change caption decoder to select essential discriminative multi-scale features to improve the change captioning performance. We evaluate our method on two RSICC datasets (e.g., LEVIR-CC and LEVIRCCD), and the experimental results demonstrate that our method achieves a state-of-the-art performance.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationRemote sensing, Dec. 2023, v. 15, no. 23, 5611en_US
dcterms.isPartOfRemote sensingen_US
dcterms.issued2023-12-
dc.identifier.scopus2-s2.0-85179131457-
dc.identifier.eissn2072-4292en_US
dc.identifier.artn5611en_US
dc.description.validate202408 bcchen_US
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumberOA_Scopus/WOS-
dc.description.fundingSourceOthersen_US
dc.description.fundingTextHong Kong Polytechnic Universityen_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryCCen_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
remotesensing-15-05611-v2.pdf2.91 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

84
Citations as of Nov 10, 2025

Downloads

56
Citations as of Nov 10, 2025

SCOPUSTM   
Citations

26
Citations as of Dec 19, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.