Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/117497
| Title: | Multi-source geo-localization in urban built environments for crowd-sourced images by contrastive learning | Authors: | Hou, Q Hou, C Zhang, F Weng, Q |
Issue Date: | Dec-2025 | Source: | ISPRS journal of photogrammetry and remote sensing, Dec. 2025, v. 230, p. 616-629 | Abstract: | Crowd-sourced images (CSIs) offer an unprecedented opportunity for gaining deeper insights into urban built environments. However, the lack of precise geographic information limits their effectiveness in various urban applications. Traditional geo-localization methods, which rely on matching CSIs with geo-tagged street-view images (SVIs), face significant challenges due to sparse coverage and temporal misalignment of reference data, especially in developing countries. To overcome these limitations, this paper proposes a novel contrastive learning framework that integrates SVIs and satellite images (SIs), utilizing a multi-scale channel attention module and InfoNCE loss to enhance the geo-localization accuracy of CSIs. Additionally, we leverage SIs to generate synthetic SVIs in areas where actual SVIs are unavailable or outdated, ensuring comprehensive coverage across diverse urban environments. A simple yet efficient data preprocessing method is proposed to align multi-view images for enhanced feature fusion. As part of our research efforts, we introduce a Multi-Source Geo-localization Dataset (MSGD) consisting of 500k geo-tagged pairs collected from 12 cities across six continents, encompassing diverse urban typologies from dense skyscraper districts to low-density areas, providing a valuable resource for future research and advancements in geo-localization methods. Our experiments show that the proposed method outperforms state-of-the-art approaches on the challenging MSGD dataset, highlighting the importance of incorporating SIs as a complementary data source for accurate geo-localization. Our code and dataset will be released at https://github.com/RCAIG/CrowdsourcingGeoLocalization. | Keywords: | Crowd-sourced images High-resolution satellite images Image geo-localization Multi-source data fusion Street-view images Urban spatial analytics |
Publisher: | Elsevier BV | Journal: | ISPRS journal of photogrammetry and remote sensing | ISSN: | 0924-2716 | EISSN: | 1872-8235 | DOI: | 10.1016/j.isprsjprs.2025.09.024 | Rights: | © 2025 The Author(s). Published by Elsevier B.V. on behalf of International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS). This is an open access article under the CC BY-NC-ND license ( http://creativecommons.org/licenses/by-nc-nd/4.0/ ). The following publication Hou, Q., Hou, C., Zhang, F., & Weng, Q. (2025). Multi-source geo-localization in urban built environments for crowd-sourced images by contrastive learning. ISPRS Journal of Photogrammetry and Remote Sensing, 230, 616-629 is available at https://doi.org/10.1016/j.isprsjprs.2025.09.024. |
| Appears in Collections: | Journal/Magazine Article |
Files in This Item:
| File | Description | Size | Format | |
|---|---|---|---|---|
| 1-s2.0-S092427162500382X-main.pdf | 7.13 MB | Adobe PDF | View/Open |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.



