Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/108219
DC FieldValueLanguage
dc.contributorDepartment of Building Environment and Energy Engineering-
dc.creatorWang, Zen_US
dc.creatorXiao, Fen_US
dc.creatorRan, Yen_US
dc.creatorLi, Yen_US
dc.creatorXu, Yen_US
dc.date.accessioned2024-07-29T02:45:59Z-
dc.date.available2024-07-29T02:45:59Z-
dc.identifier.issn0306-2619en_US
dc.identifier.urihttp://hdl.handle.net/10397/108219-
dc.language.isoenen_US
dc.publisherElsevier Ltden_US
dc.subjectEnergy costen_US
dc.subjectMulti-agent reinforcement learningen_US
dc.subjectMulti-stageen_US
dc.subjectSchedule optimizationen_US
dc.subjectThermal comforten_US
dc.titleScalable energy management approach of residential hybrid energy system using multi-agent deep reinforcement learningen_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.volume367en_US
dc.identifier.doi10.1016/j.apenergy.2024.123414en_US
dcterms.abstractDeploying renewable energy and implementing smart energy management strategies are crucial for decarbonizing Building Energy Systems (BES). Despite recent advancements in data-driven Deep Reinforcement Learning (DRL) for BES optimization, significant challenges still exist, such as the time-consuming and data-intensive nature of training DRL controllers and the complexity of environment dynamics in Multi-Agent Reinforcement Learning (MARL). Consequently, these obstacles impede the synchronization and coordination of multiple agent control, leading to slow DRL convergence performance. To address these issues. This paper proposes a novel approach to optimize hybrid building energy systems. We introduce an integrated system combining a multi-stage Proximal Policy Optimization (PPO) on-policy framework with Imitation Learning (IL), interacting with the model environment. To improve scalability and robustness of Multi-agent Systems (MAS), this approach is designed to enhance training efficiency with centralized training and decentralized execution. Simulation results of case studies demonstrate the effectiveness of the Multi-agent Deep Reinforcement Learning (MADRL) model in optimizing the operations of hybrid building energy systems in terms of indoor thermal comfort and energy efficiency. Results show the proposed framework significantly improve performance in achieving convergence in just 50 episodes for dynamic decision-making. The scalability and robustness of the proposed model have been validated across various scenarios. Compared with the baseline during cold and warm weeks, the proposed control approach achieved improvements of 34.86% and 46.10% in energy self-sufficiency ratio, respectively. Additionally, the developed MADRL effectively improved solar photovoltaic (PV) self-consumption and reduced household energy costs. Notably, it increased the average indoor temperature closer to the desired set-point by 1.33 °C, and improved the self-consumption ratio by 15.78% in the colder week and 18.47% in the warmer week, compared to baseline measurements. These findings highlight the advantages of the multi-stage PPO on-policy framework, enabling faster learning and reduced training time, resulting in cost-effective solutions and enhanced solar PV self-consumption.-
dcterms.accessRightsembargoed accessen_US
dcterms.bibliographicCitationApplied energy, 1 Aug. 2024, v. 367, 123414en_US
dcterms.isPartOfApplied energyen_US
dcterms.issued2024-08-01-
dc.identifier.scopus2-s2.0-85192857528-
dc.identifier.eissn1872-9118en_US
dc.identifier.artn123414en_US
dc.description.validate202407 bcch-
dc.identifier.FolderNumbera3093c, a3684-
dc.identifier.SubFormID49590, 50714-
dc.description.fundingSourceOthersen_US
dc.description.fundingTextthe National Natural Science Foundation of China; the Shandong Natural Science Foundation; the Xiangjiang Planen_US
dc.description.pubStatusPublisheden_US
dc.date.embargo2026-08-01en_US
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Journal/Magazine Article
Open Access Information
Status embargoed access
Embargo End Date 2026-08-01
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

58
Citations as of Apr 13, 2025

SCOPUSTM   
Citations

44
Citations as of Dec 19, 2025

WEB OF SCIENCETM
Citations

20
Citations as of Jun 5, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.