On Improving the Robustness of Reinforcement Learning-Based Controllers Using Disturbance Observer

Jeong Woo Kim1, Hyungbo Shim1, Insoon Yang1

  • 1Seoul National University

Details

10:20 - 10:40 | Wed 11 Dec | Hermes | WeA24.2

Session: Learning I

Abstract

Because reinforcement learning (RL) may cause issues in stability and safety when directly applied to physical systems, a simulator is often used to learn a control policy. However, the control performance may be easily deteriorated in a real plant due to the discrepancy between the simulator and the plant. In this paper, we propose an idea to enhance the robustness of such RL-based controllers by utilizing the disturbance observer (DOB). This method compensates for the mismatch between the plant and simulator, and rejects disturbance to maintain the nominal performance while guaranteeing robust stability. Furthermore, the proposed approach can be applied to partially observable systems. We also characterize conditions under which the learned controller has a provable performance bound when connected to the physical system.