Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/117727
| DC Field | Value | Language |
|---|---|---|
| dc.contributor | Department of Aeronautical and Aviation Engineering | - |
| dc.creator | Zhang, Y | - |
| dc.creator | Wang, Y | - |
| dc.creator | Wen, W | - |
| dc.date.accessioned | 2026-03-04T04:10:10Z | - |
| dc.date.available | 2026-03-04T04:10:10Z | - |
| dc.identifier.uri | http://hdl.handle.net/10397/117727 | - |
| dc.language.iso | en | en_US |
| dc.publisher | Institute of Electrical and Electronics Engineers | en_US |
| dc.rights | © 2025 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. | en_US |
| dc.rights | The following publication Y. Zhang, Y. Wang and W. Wen, 'EIRM-RL: Epistemic Integrity Risk Monitoring Inspired Safe Reinforcement Learning for Trustworthy Autonomous Navigation,' in IEEE Internet of Things Journal, vol. 13, no. 2, pp. 3500-3512, 15 Jan. 2026 is available at https://doi.org/10.1109/JIOT.2025.3633765. | en_US |
| dc.subject | Epistemic uncertainty | en_US |
| dc.subject | Integrity risk monitoring | en_US |
| dc.subject | Reinforcement learning (RL) | en_US |
| dc.subject | Trustworthy autonomous navigation | en_US |
| dc.subject | Unmanned ground vehicle (UGV) | en_US |
| dc.title | EIRM-RL : epistemic integrity risk monitoring inspired safe reinforcement learning for trustworthy autonomous navigation | en_US |
| dc.type | Journal/Magazine Article | en_US |
| dc.identifier.spage | 3500 | - |
| dc.identifier.epage | 3512 | - |
| dc.identifier.volume | 13 | - |
| dc.identifier.issue | 2 | - |
| dc.identifier.doi | 10.1109/JIOT.2025.3633765 | - |
| dcterms.abstract | Reinforcement learning (RL) has shown great potential for autonomous navigation within internet of things (IoT) environments, where various and changing uncertainties pose significant challenges for safe, real-world deployment. Existing safe RL methods typically employ heuristic constraints while neglecting the combined impact of multiple uncertainty sources, reducing robustness and interpretability. Drawing on concepts from global navigation satellite system (GNSS) integrity monitoring, this paper proposes an epistemic integrity risk monitoring reinforcement learning (EIRM-RL) framework to enable trustworthy autonomous navigation under uncertainty. EIRM-RL extends the GNSS protection level concept to RL by utilizing an assembled world model that quantifies and incorporates sensor noise, systematic bias, and epistemic uncertainty. Furthermore, the framework continuously monitors a dynamic epistemic risk probability, which is incorporated into policy optimization as an adaptive safety constraint via Lagrangian duality. This method enables the agent to proactively avoid hazards and effectively balance safety and performance, even in highly uncertain environments. Extensive experiments demonstrate that EIRM-RL achieves superior success rates, collision avoidance, and robustness compared to state-of-the-art safe RL methods, while maintaining high efficiency. | - |
| dcterms.accessRights | open access | en_US |
| dcterms.bibliographicCitation | IEEE internet of things journal, 15 Jan. 2026, v. 13, no. 2, p. 3500-3512 | - |
| dcterms.isPartOf | IEEE internet of things journal | - |
| dcterms.issued | 2026-01-15 | - |
| dc.identifier.scopus | 2-s2.0-105022493963 | - |
| dc.identifier.eissn | 2327-4662 | - |
| dc.description.validate | 202603 bcjz | - |
| dc.description.oa | Accepted Manuscript | en_US |
| dc.identifier.SubFormID | G001129/2026-01 | en_US |
| dc.description.fundingSource | Others | en_US |
| dc.description.fundingText | This work was supported in part by Hong Kong Innovation and Technology Fund-Innovation and Technology Support Program (ITFITSP) under the Project “Safety-Certified Multi-Source Fusion Positioning for Autonomous Vehicles in Complex Scenarios (ZPE8),” in part by the Otto Poon Charitable Foundation under the Project “Large Vision Model for UAV-UGV Collaborative Map Update (CDCG),” and in part by the Centre for Large AI Models (CLAIM) of the Hong Kong Polytechnic University. | en_US |
| dc.description.pubStatus | Published | en_US |
| dc.description.oaCategory | Green (AAM) | en_US |
| Appears in Collections: | Journal/Magazine Article | |
Files in This Item:
| File | Description | Size | Format | |
|---|---|---|---|---|
| Zhang_EIRM-RL_Epistemic_Integrity.pdf | Pre-Published version | 23.03 MB | Adobe PDF | View/Open |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.



