Adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning

Reinforcement learning provides a powerful tool for designing a satisfactory controller through interactions with the environment. Although off-policy learning algorithms were recently designed for tracking problems, most of these results either are full-state feedback or have bounded control errors...

Full description

Saved in:
Bibliographic Details
Main Authors: Chen, Ci, Xie, Lihua, Xie, Kan, Lewis, Frank L., Xie. Shengli
Other Authors: School of Electrical and Electronic Engineering
Format: Article
Language:English
Published: 2022
Subjects:
Online Access:https://hdl.handle.net/10356/163293
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-163293
record_format dspace
spelling sg-ntu-dr.10356-1632932022-11-30T06:22:35Z Adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning Chen, Ci Xie, Lihua Xie, Kan Lewis, Frank L. Xie. Shengli School of Electrical and Electronic Engineering Engineering::Electrical and electronic engineering Reinforcement Learning Output Tracking Reinforcement learning provides a powerful tool for designing a satisfactory controller through interactions with the environment. Although off-policy learning algorithms were recently designed for tracking problems, most of these results either are full-state feedback or have bounded control errors, which may not be flexible or desirable for engineering problems in the real world. To address these problems, we propose an output-feedback-based reinforcement learning approach that allows us to find the optimal control solution using input–output data and ensure the asymptotic tracking control of continuous-time systems. More specifically, we first propose a dynamical controller revised from the standard output regulation theory and use it to formulate an optimal output tracking problem. Then, a state observer is used to re-express the system state. Consequently, we address the rank issue of the parameterization matrix and analyze the state re-expression error that are crucial for transforming the off-policy learning into an output-feedback form. A comprehensive simulation study is given to demonstrate the effectiveness of the proposed approach. 2022-11-30T06:22:35Z 2022-11-30T06:22:35Z 2022 Journal Article Chen, C., Xie, L., Xie, K., Lewis, F. L. & Xie. Shengli (2022). Adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning. Automatica, 146, 110581-. https://dx.doi.org/10.1016/j.automatica.2022.110581 0005-1098 https://hdl.handle.net/10356/163293 10.1016/j.automatica.2022.110581 2-s2.0-85137619440 146 110581 en Automatica © 2022 Elsevier Ltd. All rights reserved.
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering
Reinforcement Learning
Output Tracking
spellingShingle Engineering::Electrical and electronic engineering
Reinforcement Learning
Output Tracking
Chen, Ci
Xie, Lihua
Xie, Kan
Lewis, Frank L.
Xie. Shengli
Adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning
description Reinforcement learning provides a powerful tool for designing a satisfactory controller through interactions with the environment. Although off-policy learning algorithms were recently designed for tracking problems, most of these results either are full-state feedback or have bounded control errors, which may not be flexible or desirable for engineering problems in the real world. To address these problems, we propose an output-feedback-based reinforcement learning approach that allows us to find the optimal control solution using input–output data and ensure the asymptotic tracking control of continuous-time systems. More specifically, we first propose a dynamical controller revised from the standard output regulation theory and use it to formulate an optimal output tracking problem. Then, a state observer is used to re-express the system state. Consequently, we address the rank issue of the parameterization matrix and analyze the state re-expression error that are crucial for transforming the off-policy learning into an output-feedback form. A comprehensive simulation study is given to demonstrate the effectiveness of the proposed approach.
author2 School of Electrical and Electronic Engineering
author_facet School of Electrical and Electronic Engineering
Chen, Ci
Xie, Lihua
Xie, Kan
Lewis, Frank L.
Xie. Shengli
format Article
author Chen, Ci
Xie, Lihua
Xie, Kan
Lewis, Frank L.
Xie. Shengli
author_sort Chen, Ci
title Adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning
title_short Adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning
title_full Adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning
title_fullStr Adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning
title_full_unstemmed Adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning
title_sort adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning
publishDate 2022
url https://hdl.handle.net/10356/163293
_version_ 1751548570401505280