Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/117497
PIRA download icon_1.1View/Download Full Text
Title: Multi-source geo-localization in urban built environments for crowd-sourced images by contrastive learning
Authors: Hou, Q 
Hou, C
Zhang, F
Weng, Q 
Issue Date: Dec-2025
Source: ISPRS journal of photogrammetry and remote sensing, Dec. 2025, v. 230, p. 616-629
Abstract: Crowd-sourced images (CSIs) offer an unprecedented opportunity for gaining deeper insights into urban built environments. However, the lack of precise geographic information limits their effectiveness in various urban applications. Traditional geo-localization methods, which rely on matching CSIs with geo-tagged street-view images (SVIs), face significant challenges due to sparse coverage and temporal misalignment of reference data, especially in developing countries. To overcome these limitations, this paper proposes a novel contrastive learning framework that integrates SVIs and satellite images (SIs), utilizing a multi-scale channel attention module and InfoNCE loss to enhance the geo-localization accuracy of CSIs. Additionally, we leverage SIs to generate synthetic SVIs in areas where actual SVIs are unavailable or outdated, ensuring comprehensive coverage across diverse urban environments. A simple yet efficient data preprocessing method is proposed to align multi-view images for enhanced feature fusion. As part of our research efforts, we introduce a Multi-Source Geo-localization Dataset (MSGD) consisting of 500k geo-tagged pairs collected from 12 cities across six continents, encompassing diverse urban typologies from dense skyscraper districts to low-density areas, providing a valuable resource for future research and advancements in geo-localization methods. Our experiments show that the proposed method outperforms state-of-the-art approaches on the challenging MSGD dataset, highlighting the importance of incorporating SIs as a complementary data source for accurate geo-localization. Our code and dataset will be released at https://github.com/RCAIG/CrowdsourcingGeoLocalization.
Keywords: Crowd-sourced images
High-resolution satellite images
Image geo-localization
Multi-source data fusion
Street-view images
Urban spatial analytics
Publisher: Elsevier BV
Journal: ISPRS journal of photogrammetry and remote sensing 
ISSN: 0924-2716
EISSN: 1872-8235
DOI: 10.1016/j.isprsjprs.2025.09.024
Rights: © 2025 The Author(s). Published by Elsevier B.V. on behalf of International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS). This is an open access article under the CC BY-NC-ND license ( http://creativecommons.org/licenses/by-nc-nd/4.0/ ).
The following publication Hou, Q., Hou, C., Zhang, F., & Weng, Q. (2025). Multi-source geo-localization in urban built environments for crowd-sourced images by contrastive learning. ISPRS Journal of Photogrammetry and Remote Sensing, 230, 616-629 is available at https://doi.org/10.1016/j.isprsjprs.2025.09.024.
Appears in Collections:Journal/Magazine Article

Files in This Item:
File Description SizeFormat 
1-s2.0-S092427162500382X-main.pdf7.13 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show full item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.