Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/115707
DC FieldValueLanguage
dc.contributorDepartment of Electrical and Electronic Engineeringen_US
dc.creatorTang, Jen_US
dc.creatorDuan, Xen_US
dc.creatorZhou, Jen_US
dc.creatorQu, Ken_US
dc.creatorHo, IWHen_US
dc.creatorTian, Den_US
dc.date.accessioned2025-10-23T06:29:09Z-
dc.date.available2025-10-23T06:29:09Z-
dc.identifier.issn0018-9545en_US
dc.identifier.urihttp://hdl.handle.net/10397/115707-
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineersen_US
dc.subjectData collectionen_US
dc.subjectDeep reinforcement learningen_US
dc.subjectEnergy efficienten_US
dc.subjectPath planningen_US
dc.subjectTask offloadingen_US
dc.subjectUAVen_US
dc.titleEnergy-efficient data collection and task offloading optimization in heterogeneous multi-tier UAV systems via deep reinforcement learningen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.doi10.1109/TVT.2025.3622980en_US
dcterms.abstractUnmanned aerial vehicles (UAVs) have gained considerable attention in data collection due to their mobility and flexibility. These capabilities are crucial in time-sensitive missions (e.g., disaster response, military reconnaissance). In such cases, tasks are often subject to tight deadlines and require timely access to information. To address these challenges, this paper investigates collaborative data collection and task offloading in multi-UAV systems, aiming to maximize mission area coverage while minimizing total energy consumption. To overcome the limited computing power of data collection UAVs, we propose a heterogeneous multi-tier UAV system. In this design, an assisted UAV with strong computing capabilities is introduced to handle data offloading and processing. This enhances energy efficiency and enables timely task execution. Consequently, we develop an integrated optimization model to jointly design trajectory planning and task offloading under communication, energy, and deadline constraints. We propose a deep reinforcement learning algorithm called data collection optimized proximal policy optimization (DCOPPO). This approach optimizes both UAV trajectories and offloading decisions. Simulation results demonstrate that DCOPPO significantly outperforms baseline DRL approaches in terms of energy efficiency and task completion performance.en_US
dcterms.accessRightsembargoed accessen_US
dcterms.bibliographicCitationIEEE transactions on vehicular technology, Date of Publication: 17 October 2025, Early Access, https://doi.org/10.1109/TVT.2025.3622980en_US
dcterms.isPartOfIEEE transactions on vehicular technologyen_US
dcterms.issued2025-
dc.identifier.eissn1939-9359en_US
dc.description.validate202510 bcchen_US
dc.description.oaNot applicableen_US
dc.identifier.FolderNumbera4136-
dc.identifier.SubFormID52129-
dc.description.fundingSourceOthersen_US
dc.description.fundingTextThis work was supported by the National Natural Science Foundation of China under Grant 62173012, Grant 62432002, Grant U2433202 and Grant U22A2046, the Fundamental Research Funds for the Central Universities (Beihang Ganwei Action Plan Key Program) under Grant JK2024-19.en_US
dc.description.pubStatusEarly releaseen_US
dc.date.embargo0000-00-00 (to be updated)en_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Journal/Magazine Article
Open Access Information
Status embargoed access
Embargo End Date 0000-00-00 (to be updated)
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.