Safe Intermittent Reinforcement Learning for Nonlinear Systems

Yongliang Yang1, Kyriakos G. Vamvoudakis2, Hamidreza Modares3, Wei He1, Yi-xin Yin1, Donald Wunsch4

  • 1University of Science and Technology Beijing
  • 2Georgia Inst. of Tech
  • 3Michigan State University
  • 4University of Missouri-Rolla

Details

10:00 - 10:20 | Wed 11 Dec | Rhodes 10 | WeA20.1

Session: Event-Triggered and Self-Triggered Control Based on Optimization Methods

Abstract

In this paper, an online intermittent actor-critic reinforcement learning method is used to stabilize nonlinear systems optimally while also guaranteeing safety. A barrier function-based transformation is introduced to ensure that the system does not violate the user-defined safety constraints. It is shown that the safety constraints of the original system can be guaranteed by assuring the stability of the equilibrium point of an appropriately transformed system. Then, an online intermittent actor-critic learning framework is developed to learn the optimal safe intermittent controller. Also, Zeno behavior is guaranteed to be excluded. Finally, numerical examples are conducted to verify the efficacy of the learning algorithm.