Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/112855
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Applied Mathematics-
dc.creatorHe, Y-
dc.creatorHu, R-
dc.creatorLiang, K-
dc.creatorLiu, Y-
dc.creatorZhou, Z-
dc.date.accessioned2025-05-09T06:12:43Z-
dc.date.available2025-05-09T06:12:43Z-
dc.identifier.urihttp://hdl.handle.net/10397/112855-
dc.language.isoenen_US
dc.publisherMDPI AGen_US
dc.rightsCopyright: © 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).en_US
dc.rightsThe following publication He, Y., Hu, R., Liang, K., Liu, Y., & Zhou, Z. (2025). Deep Reinforcement Learning Algorithm with Long Short-Term Memory Network for Optimizing Unmanned Aerial Vehicle Information Transmission. Mathematics, 13(1), 46 is available at https://doi.org/10.3390/math13010046.en_US
dc.subjectDeep reinforcement learning (DRL)en_US
dc.subjectLong short-term memory (LSTM)en_US
dc.subjectNonconvex optimizationen_US
dc.subjectOptimal controlen_US
dc.subjectUnmanned aerial vehicle (UAV)en_US
dc.titleDeep reinforcement learning algorithm with long short-term memory network for optimizing unmanned aerial vehicle information transmissionen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume13-
dc.identifier.issue1-
dc.identifier.doi10.3390/math13010046-
dcterms.abstractThe optimization of information transmission in unmanned aerial vehicles (UAVs) is essential for enhancing their operational efficiency across various applications. This issue is framed as a mixed-integer nonconvex optimization challenge, which traditional optimization algorithms and reinforcement learning (RL) methods often struggle to address effectively. In this paper, we propose a novel deep reinforcement learning algorithm that utilizes a hybrid discrete–continuous action space. To address the long-term dependency issues inherent in UAV operations, we incorporate a long short-term memory (LSTM) network. Our approach accounts for the specific flight constraints of fixed-wing UAVs and employs a continuous policy network to facilitate real-time flight path planning. A non-sparse reward function is designed to maximize data collection from internet of things (IoT) devices, thus guiding the UAV to optimize its operational efficiency. Experimental results demonstrate that the proposed algorithm yields near-optimal flight paths and significantly improves data collection capabilities, compared to conventional heuristic methods, achieving an improvement of up to 10.76%. Validation through simulations confirms the effectiveness and practicality of the proposed approach in real-world scenarios.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationMathematics, Jan. 2025, v. 13, no. 1, 46-
dcterms.isPartOfMathematics-
dcterms.issued2025-01-
dc.identifier.scopus2-s2.0-85214506518-
dc.identifier.eissn2227-7390-
dc.identifier.artn46-
dc.description.validate202505 bcch-
dc.description.oaVersion of Recorden_US
dc.identifier.FolderNumberOA_Scopus/WOSen_US
dc.description.fundingSourceOthersen_US
dc.description.fundingTextThe National Natural Science Foundation of China (No. 42171412)en_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryCCen_US
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
mathematics-13-00046.pdf1.75 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

SCOPUSTM   
Citations

6
Citations as of Dec 19, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.