Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/101902
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computing-
dc.creatorLiang, Z-
dc.creatorCao, J-
dc.creatorJiang, S-
dc.creatorSaxena, D-
dc.creatorXu, H-
dc.date.accessioned2023-09-22T06:58:31Z-
dc.date.available2023-09-22T06:58:31Z-
dc.identifier.isbn978-1-6654-7177-0-
dc.identifier.urihttp://hdl.handle.net/10397/101902-
dc.description42nd IEEE International Conference on Distributed Computing Systems, July 10 - July 13, 2022, Bologna, Italyen_US
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineers Inc.en_US
dc.rights© 2022 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.en_US
dc.rightsThe following publication Z. Liang, J. Cao, S. Jiang, D. Saxena and H. Xu, "Hierarchical Reinforcement Learning with Opponent Modeling for Distributed Multi-agent Cooperation," 2022 IEEE 42nd International Conference on Distributed Computing Systems (ICDCS), Bologna, Italy, 2022, pp. 884-894 is available at https://doi.org/10.1109/ICDCS54860.2022.00090.en_US
dc.subjectDeep Reinforcement Learningen_US
dc.subjectHierarchical Reinforcement Learningen_US
dc.subjectMulti-agent Cooperationen_US
dc.titleHierarchical reinforcement learning with opponent modeling for distributed multi-agent cooperationen_US
dc.typeConference Paperen_US
dc.identifier.spage884-
dc.identifier.epage894-
dc.identifier.doi10.1109/ICDCS54860.2022.00090-
dcterms.abstractMany real-world applications can be formulated as multi-agent cooperation problems, such as network packet routing and coordination of autonomous vehicles. The emergence of deep reinforcement learning (DRL) provides a promising approach for multi-agent cooperation through the interaction of the agents and environments. However, traditional DRL solutions suffer from the high dimensions of multiple agents with continuous action space during policy search. Besides, the dynamicity of agents’ policies makes the training non-stationary. To tackle the issues, we propose a hierarchical reinforcement learning approach with high-level decision-making and low-level individual control for efficient policy search. In particular, the cooperation of multiple agents can be learned in high-level discrete action space efficiently. At the same time, the low-level individual control can be reduced to single-agent reinforcement learning. In addition to hierarchical reinforcement learning, we propose an opponent modeling network to model other agents’ policies during the learning process. In contrast to end-to-end DRL approaches, our approach reduces the learning complexity by decomposing the overall task into sub-tasks in a hierarchical way. To evaluate the efficiency of our approach, we conduct a real-world case study in the cooperative lane change scenario. Both simulation and real-world experiments show the superiority of our approach in the collision rate and convergence speed.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationProceedings: 2022 IEEE 42nd International Conference on Distributed Computing Systems ICDCS 2022 : Bologna, Italy, 10-13 July 2022, p. 884-894-
dcterms.issued2022-
dc.identifier.scopus2-s2.0-85140909443-
dc.relation.conferenceInternational Conference on Distributed Computing Systems [ICDCS]-
dc.description.validate202309 bcch-
dc.description.oaAccepted Manuscripten_US
dc.identifier.FolderNumbera2428en_US
dc.identifier.SubFormID47664en_US
dc.description.fundingSourceRGCen_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
IEEE_HERO_for_Distributed_Multi_agent_Cooperation.pdfPre-Published version2.61 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

115
Citations as of Apr 14, 2025

Downloads

137
Citations as of Apr 14, 2025

SCOPUSTM   
Citations

17
Citations as of Dec 19, 2025

WEB OF SCIENCETM
Citations

5
Citations as of Oct 10, 2024

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.