Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/115768
| DC Field | Value | Language |
|---|---|---|
| dc.contributor | Department of Building Environment and Energy Engineering | en_US |
| dc.creator | Wang, S | en_US |
| dc.creator | Zhao, H | en_US |
| dc.creator | Shu, T | en_US |
| dc.creator | Pan, Z | en_US |
| dc.creator | Liang, G | en_US |
| dc.creator | Zhao, J | en_US |
| dc.date.accessioned | 2025-10-28T07:33:07Z | - |
| dc.date.available | 2025-10-28T07:33:07Z | - |
| dc.identifier.issn | 1949-3053 | en_US |
| dc.identifier.uri | http://hdl.handle.net/10397/115768 | - |
| dc.language.iso | en | en_US |
| dc.publisher | Institute of Electrical and Electronics Engineers | en_US |
| dc.rights | © 2025 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. | en_US |
| dc.rights | The following publication S. Wang, H. Zhao, T. Shu, Z. Pan, G. Liang and J. Zhao, 'Power Smoothing Control for Wind-Storage Integrated Systems With Hierarchical Safe Reinforcement Learning and Curriculum Learning,' in IEEE Transactions on Smart Grid, vol. 16, no. 6, pp. 4606-4619, Nov. 2025 is available at https://doi.org/10.1109/TSG.2025.3593340. | en_US |
| dc.subject | Curriculum learning | en_US |
| dc.subject | Deep reinforcement learning | en_US |
| dc.subject | Power smoothing control | en_US |
| dc.subject | Prioritized experiment replay | en_US |
| dc.subject | Wind storage integrated systems | en_US |
| dc.title | Power smoothing control for wind-storage integrated systems with hierarchical safe reinforcement learning and curriculum learning | en_US |
| dc.type | Journal/Magazine Article | en_US |
| dc.identifier.spage | 4606 | en_US |
| dc.identifier.epage | 4619 | en_US |
| dc.identifier.volume | 16 | en_US |
| dc.identifier.issue | 6 | en_US |
| dc.identifier.doi | 10.1109/TSG.2025.3593340 | en_US |
| dcterms.abstract | As the penetration of wind energy increases, the Wind Storage Integrated System (WSIS) has become a critical solution to ensure stable wind power output and maximize economic benefits. However, existing data-based power smoothing control strategies are hard to satisfy the power fluctuation constraint in a complex environment, presenting an inefficient coordinate performance. To address this problem, this paper proposes a novel Hierarchical Safe Deep Reinforcement Learning (HSDRL) control framework for WSIS. The control problem is first reformulated as two interconnected Constrained Markov Decision Processes, and the hierarchical primal-dual-based safe Deep Deterministic Policy Gradient algorithm is proposed to learn the optimal policy that ensures the power output constraint. Furthermore, the curriculum learning is designed and Constraint Violation Prioritized Experience Replay method is proposed to address the unstable convergence issues caused by imbalanced constraint violation and constraint satisfaction experience data. Last, a hierarchical shared feature neural network structure is designed to share the parameters of Q networks at hierarchies and increase learning efficiency. Simulation results in WindFarmSimulator validate the efficacy of the proposed control framework, demonstrating a 15.3% improvement in profit and a 46.0% reduction in fluctuation compared to existing methods. | en_US |
| dcterms.accessRights | open access | en_US |
| dcterms.bibliographicCitation | IEEE transactions on smart grid, Nov. 2025, v. 16, no. 6, p. 4606-4619 | en_US |
| dcterms.isPartOf | IEEE transactions on smart grid | en_US |
| dcterms.issued | 2025-11 | - |
| dc.identifier.scopus | 2-s2.0-105012117574 | - |
| dc.identifier.eissn | 1949-3061 | en_US |
| dc.description.validate | 202510 bcjz | en_US |
| dc.description.oa | Accepted Manuscript | en_US |
| dc.identifier.SubFormID | G000298/2025-08 | - |
| dc.description.fundingSource | Others | en_US |
| dc.description.fundingText | This work was supported in part by the National Natural Science Foundation of China under Grant 72331009, Grant 72171206, and Grant 92270105; in part by the Shenzhen Key Lab of Crowd Intelligence Empowered Low-Carbon Energy Network under Grant ZDSYS20220606100601002; in part by the Guangdong Power Grid Company under Grant GDKJXM20231024; in part by the PolyU Direct under Grant P0047700, Grant P0043885, and Grant P0051105; in part by the Shenzhen Natural Science Fund; in part by the Stable Support Plan Program under Grant GXWD20231128112434001; and in part by the Shenzhen Institute of Artificial Intelligence and Robotics for Society. Paper no. TSG-02300-2024 | en_US |
| dc.description.pubStatus | Published | en_US |
| dc.description.oaCategory | Green (AAM) | en_US |
| Appears in Collections: | Journal/Magazine Article | |
Files in This Item:
| File | Description | Size | Format | |
|---|---|---|---|---|
| Wang_Power_Smoothing_Control.pdf | Pre-Published version | 2.28 MB | Adobe PDF | View/Open |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.



