Adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning
Reinforcement learning provides a powerful tool for designing a satisfactory controller through interactions with the environment. Although off-policy learning algorithms were recently designed for tracking problems, most of these results either are full-state feedback or have bounded control errors...
Saved in:
Main Authors: | , , , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2022
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/163293 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-163293 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1632932022-11-30T06:22:35Z Adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning Chen, Ci Xie, Lihua Xie, Kan Lewis, Frank L. Xie. Shengli School of Electrical and Electronic Engineering Engineering::Electrical and electronic engineering Reinforcement Learning Output Tracking Reinforcement learning provides a powerful tool for designing a satisfactory controller through interactions with the environment. Although off-policy learning algorithms were recently designed for tracking problems, most of these results either are full-state feedback or have bounded control errors, which may not be flexible or desirable for engineering problems in the real world. To address these problems, we propose an output-feedback-based reinforcement learning approach that allows us to find the optimal control solution using input–output data and ensure the asymptotic tracking control of continuous-time systems. More specifically, we first propose a dynamical controller revised from the standard output regulation theory and use it to formulate an optimal output tracking problem. Then, a state observer is used to re-express the system state. Consequently, we address the rank issue of the parameterization matrix and analyze the state re-expression error that are crucial for transforming the off-policy learning into an output-feedback form. A comprehensive simulation study is given to demonstrate the effectiveness of the proposed approach. 2022-11-30T06:22:35Z 2022-11-30T06:22:35Z 2022 Journal Article Chen, C., Xie, L., Xie, K., Lewis, F. L. & Xie. Shengli (2022). Adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning. Automatica, 146, 110581-. https://dx.doi.org/10.1016/j.automatica.2022.110581 0005-1098 https://hdl.handle.net/10356/163293 10.1016/j.automatica.2022.110581 2-s2.0-85137619440 146 110581 en Automatica © 2022 Elsevier Ltd. All rights reserved. |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Electrical and electronic engineering Reinforcement Learning Output Tracking |
spellingShingle |
Engineering::Electrical and electronic engineering Reinforcement Learning Output Tracking Chen, Ci Xie, Lihua Xie, Kan Lewis, Frank L. Xie. Shengli Adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning |
description |
Reinforcement learning provides a powerful tool for designing a satisfactory controller through interactions with the environment. Although off-policy learning algorithms were recently designed for tracking problems, most of these results either are full-state feedback or have bounded control errors, which may not be flexible or desirable for engineering problems in the real world. To address these problems, we propose an output-feedback-based reinforcement learning approach that allows us to find the optimal control solution using input–output data and ensure the asymptotic tracking control of continuous-time systems. More specifically, we first propose a dynamical controller revised from the standard output regulation theory and use it to formulate an optimal output tracking problem. Then, a state observer is used to re-express the system state. Consequently, we address the rank issue of the parameterization matrix and analyze the state re-expression error that are crucial for transforming the off-policy learning into an output-feedback form. A comprehensive simulation study is given to demonstrate the effectiveness of the proposed approach. |
author2 |
School of Electrical and Electronic Engineering |
author_facet |
School of Electrical and Electronic Engineering Chen, Ci Xie, Lihua Xie, Kan Lewis, Frank L. Xie. Shengli |
format |
Article |
author |
Chen, Ci Xie, Lihua Xie, Kan Lewis, Frank L. Xie. Shengli |
author_sort |
Chen, Ci |
title |
Adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning |
title_short |
Adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning |
title_full |
Adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning |
title_fullStr |
Adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning |
title_full_unstemmed |
Adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning |
title_sort |
adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning |
publishDate |
2022 |
url |
https://hdl.handle.net/10356/163293 |
_version_ |
1751548570401505280 |