Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/116858
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Land Surveying and Geo-Informatics-
dc.creatorWan, Q-
dc.creatorHe, M-
dc.creatorLiu, Z-
dc.creatorZhang, S-
dc.creatorChen, L-
dc.creatorChen, R-
dc.creatorYu, Y-
dc.date.accessioned2026-01-21T03:53:23Z-
dc.date.available2026-01-21T03:53:23Z-
dc.identifier.issn1569-8432-
dc.identifier.urihttp://hdl.handle.net/10397/116858-
dc.language.isoenen_US
dc.publisherElsevier BVen_US
dc.rights© 2025 The Author(s). Published by Elsevier B.V. This is an open access article under the CC BY license ( http://creativecommons.org/licenses/by/4.0/ ).en_US
dc.rightsThe following publication Wan, Q., He, M., Liu, Z., Zhang, S., Chen, L., Chen, R., & Yu, Y. (2025). UMTrajNet: A spatiotemporal uncertainty modelling structure of human movement trajectories under complex indoor areas. International Journal of Applied Earth Observation and Geoinformation, 143, 104780 is available at https://doi.org/10.1016/j.jag.2025.104780.en_US
dc.subjectGCNen_US
dc.subjectIndoor human movement trajectoryen_US
dc.subjectLSTMen_US
dc.subjectMovement uncertainty modelingen_US
dc.subjectSpatiotemporal structureen_US
dc.titleUMTrajNet : a spatiotemporal uncertainty modelling structure of human movement trajectories under complex indoor areasen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume143-
dc.identifier.doi10.1016/j.jag.2025.104780-
dcterms.abstractMovement uncertainty modeling is essential for reliable human mobility analytics, especially in indoor spaces with denied Global Navigation Satellite System (GNSS) signals and complex human motion modes. However, the performance of current uncertainty modelling approaches is hindered by the dynamic spatial–temporal correlations within the indoor human movement trajectory. To tackle these problems, this paper proposes a novel uncertainty modelling framework UMTrajNet to address indoor human movement reconstruction, movement uncertainty error prediction, and uncertainty region modelling. A spatiotemporal structure is designed by combining the Graph Convolutional Networks (GCN) network and the Long Short-Term Memory (LSTM) network, enabling adaptive learning the spatial and temporal features of human movement. An adaptive Euclidean distance method is employed to generate continuous uncertainty regions for the selected indoor trajectory based on extracted spatiotemporal features. By integrating GCN with LSTM, UMTrajNet effectively captures the complex spatial–temporal dependencies within the trajectory data thereby overcoming the limitations of traditional solely LSTM-based models in spatial context correlation modeling. Real-world experimental results indicate that the presented UMTrajNet significantly outperforms the state-of-the-art baseline models in uncertainty modelling across different datasets.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationInternational journal of applied earth observation and geoinformation, Sept 2025, v. 143, 104780-
dcterms.isPartOfInternational journal of applied earth observation and geoinformation-
dcterms.issued2025-09-
dc.identifier.scopus2-s2.0-105012839864-
dc.identifier.eissn1872-826X-
dc.identifier.artn104780-
dc.description.validate202601 bcch-
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumberOA_Scopus/WOSen_US
dc.description.fundingSourceOthersen_US
dc.description.fundingTextThis work was supported by the Hong Kong Polytechnic University (P0045937); Open Fund of State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University (Grant No. 23P03).en_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryCCen_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
1-s2.0-S1569843225004273-main.pdf6.22 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.