Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/95942
| DC Field | Value | Language |
|---|---|---|
| dc.contributor | Department of Applied Mathematics | en_US |
| dc.creator | Li, N | en_US |
| dc.creator | Li, X | en_US |
| dc.creator | Peng, J | en_US |
| dc.creator | Xu, ZQ | en_US |
| dc.date.accessioned | 2022-10-28T07:28:23Z | - |
| dc.date.available | 2022-10-28T07:28:23Z | - |
| dc.identifier.issn | 0018-9286 | en_US |
| dc.identifier.uri | http://hdl.handle.net/10397/95942 | - |
| dc.language.iso | en | en_US |
| dc.publisher | Institute of Electrical and Electronics Engineers | en_US |
| dc.rights | © 2022 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. | en_US |
| dc.rights | The following publication N. Li, X. Li, J. Peng and Z. Q. Xu, "Stochastic Linear Quadratic Optimal Control Problem: A Reinforcement Learning Method," in IEEE Transactions on Automatic Control, vol. 67, no. 9, pp. 5009-5016, Sept. 2022 is available at https://dx.doi.org/10.1109/TAC.2022.3181248. | en_US |
| dc.subject | Optimal control | en_US |
| dc.subject | Stochastic processes | en_US |
| dc.subject | Heuristic algorithms | en_US |
| dc.subject | Trajectory | en_US |
| dc.subject | Mathematics | en_US |
| dc.subject | Mathematical models | en_US |
| dc.subject | Riccati equations | en_US |
| dc.title | Stochastic linear quadratic optimal control problem : a reinforcement learning method | en_US |
| dc.type | Journal/Magazine Article | en_US |
| dc.identifier.spage | 5009 | en_US |
| dc.identifier.epage | 5016 | en_US |
| dc.identifier.volume | 67 | en_US |
| dc.identifier.issue | 9 | en_US |
| dc.identifier.doi | 10.1109/TAC.2022.3181248 | en_US |
| dcterms.abstract | This paper adopts a reinforcement learning (RL) method to solve infinite horizon continuous-time stochastic linear quadratic problems, where the drift and diffusion terms in the dynamics may depend on both the state and control. Based on Bellman’s dynamic programming principle, we present an online RL algorithm to attain optimal control with partial system information. This algorithm computes the optimal control rather than estimates the system coefficients and solves the related Riccati equation. It only requires local trajectory information, which significantly simplifies the calculation process. We shed light on our theoretical findings using two numerical examples. | en_US |
| dcterms.accessRights | open access | en_US |
| dcterms.bibliographicCitation | IEEE transactions on automatic control, Sept 2022, v. 67, no. 9, p. 5009-5016 | en_US |
| dcterms.isPartOf | IEEE transactions on automatic control | en_US |
| dcterms.issued | 2022-09 | - |
| dc.identifier.eissn | 1558-2523 | en_US |
| dc.description.validate | 202207 bcch | en_US |
| dc.description.oa | Accepted Manuscript | en_US |
| dc.identifier.FolderNumber | a1453, a2099 | - |
| dc.identifier.SubFormID | 45032, 46594 | - |
| dc.description.fundingSource | RGC | en_US |
| dc.description.pubStatus | Published | en_US |
| dc.description.oaCategory | Green (AAM) | en_US |
| Appears in Collections: | Journal/Magazine Article | |
Files in This Item:
| File | Description | Size | Format | |
|---|---|---|---|---|
| Li_Stochastic_Linear_Quadratic.pdf | Pre-Published version | 1.43 MB | Adobe PDF | View/Open |
Page views
91
Last Week
0
0
Last month
Citations as of Apr 14, 2025
Downloads
214
Citations as of Apr 14, 2025
SCOPUSTM
Citations
10
Citations as of Jun 21, 2024
WEB OF SCIENCETM
Citations
12
Citations as of Oct 10, 2024
Google ScholarTM
Check
Altmetric
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.



