Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/118121
| DC Field | Value | Language |
|---|---|---|
| dc.contributor | Department of Aeronautical and Aviation Engineering | - |
| dc.creator | Zhang, Y | - |
| dc.creator | Wang, Y | - |
| dc.creator | Yan, P | - |
| dc.creator | Wen, W | - |
| dc.date.accessioned | 2026-03-18T02:30:14Z | - |
| dc.date.available | 2026-03-18T02:30:14Z | - |
| dc.identifier.issn | 1524-9050 | - |
| dc.identifier.uri | http://hdl.handle.net/10397/118121 | - |
| dc.language.iso | en | en_US |
| dc.publisher | Institute of Electrical and Electronics Engineers | en_US |
| dc.rights | © 2025 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. | en_US |
| dc.rights | The following publication Y. Zhang, Y. Wang, P. Yan and W. Wen, 'Learning Safe, Optimal, and Real-Time Flight Interaction With Deep Confidence-Enhanced Reachability Guarantee,' in IEEE Transactions on Intelligent Transportation Systems, vol. 26, no. 12, pp. 23100-23112, Dec. 2025 is available at https://doi.org/10.1109/TITS.2025.3616580. | en_US |
| dc.subject | Deep confidence-enhanced reachability guarantees | en_US |
| dc.subject | Deep reinforcement learning | en_US |
| dc.subject | Joint planning and control | en_US |
| dc.subject | Unmanned aerial vehicles | en_US |
| dc.title | Learning safe, optimal, and real-time flight interaction with deep confidence-enhanced reachability guarantee | en_US |
| dc.type | Journal/Magazine Article | en_US |
| dc.identifier.spage | 23100 | - |
| dc.identifier.epage | 23112 | - |
| dc.identifier.volume | 26 | - |
| dc.identifier.issue | 12 | - |
| dc.identifier.doi | 10.1109/TITS.2025.3616580 | - |
| dcterms.abstract | In the low-altitude economy, ensuring the safe and agile flight of unmanned aerial vehicles (UAVs) in dynamic obstacle environments is essential for expanding interactive applications like parcel delivery. While deep reinforcement learning (DRL) shows promise for UAV motion planning and control, its trial-and-error exploration often struggles to ensure both agility and safety, especially under uncertain observational noise. Therefore, this paper proposes a deep confidence-enhanced reachability policy optimization (DCRPO) framework. By integrating safe DRL with nonlinear model predictive control (NMPC), DCRPO achieves high-level safety decisions, complex real-time joint planning and control for UAVs. Furthermore, we develop a deep confidence-enhanced reachability guarantee that constructs a set of stochastically forward-reachable planned trajectories under uncertainty, enabling robust safety collision probability certifications. This safe reachability mechanism adaptively selects belief space actions from planned actions to interact with the environment, further enhancing safety and reducing training time. In extensive experiments of UAVs traversing a fast-moving rectangular gate, the proposed method outperforms other state-of-the-art baseline methods under varying environments in terms of operational robustness. Furthermore, the proposed method significantly reduces overall collision violations and training time, greatly improving both training safety and efficiency. The demonstration video (https://youtu.be/7xkp9U7FSJg) and the source code (https://github.com/ZyyFLY/DCRPO) are also provided. | - |
| dcterms.accessRights | open access | en_US |
| dcterms.bibliographicCitation | IEEE transactions on intelligent transportation systems, Dec. 2025, v. 26, no. 12, p. 23100-23112 | - |
| dcterms.isPartOf | IEEE transactions on intelligent transportation systems | - |
| dcterms.issued | 2025-12 | - |
| dc.identifier.scopus | 2-s2.0-105019099058 | - |
| dc.identifier.eissn | 1558-0016 | - |
| dc.description.validate | 202603 bcjz | - |
| dc.description.oa | Accepted Manuscript | en_US |
| dc.identifier.SubFormID | G001270/2026-02 | en_US |
| dc.description.fundingSource | Others | en_US |
| dc.description.fundingText | This work was supported in part by Hong Kong Innovation and Technology Fund-Innovation and Technology Support Program (ITF-ITSP) under the Project “Safety-Certified Multi-Source Fusion Positioning for Autonomous Vehicles in Complex Scenarios (ZPE8).” | en_US |
| dc.description.pubStatus | Published | en_US |
| dc.description.oaCategory | Green (AAM) | en_US |
| Appears in Collections: | Journal/Magazine Article | |
Files in This Item:
| File | Description | Size | Format | |
|---|---|---|---|---|
| Zhang_Learning_Safe_Optimal.pdf | Pre-Published version | 20.18 MB | Adobe PDF | View/Open |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.



