TY - JOUR
T1 - Improving the Robustness of Reinforcement Learning Policies With L1Adaptive Control
AU - Cheng, Yikun
AU - Zhao, Pan
AU - Wang, Fanxin
AU - Block, Daniel J.
AU - Hovakimyan, Naira
N1 - Publisher Copyright:
© 2016 IEEE.
PY - 2022/7
Y1 - 2022/7
N2 - A reinforcement learning (RL) control policy could fail in a new/perturbed environment that is different from the training environment, due to the presence of dynamic variations. For controlling systems with continuous state and action spaces, we propose an add-on approach to robustifying a pre-trained RL policy by augmenting it with an ${\mathcal {L}_{1}}$ adaptive controller (${\mathcal {L}_{1}}$AC). Leveraging the capability of an ${\mathcal {L}_{1}}$AC for fast estimation and active compensation of dynamic variations, the proposed approach can improve the robustness of an RL policy which is trained either in a simulator or in the real world without consideration of a broad class of dynamic variations. Numerical and real-world experiments empirically demonstrate the efficacy of the proposed approach in robustifying RL policies trained using both model-free and model-based methods.
AB - A reinforcement learning (RL) control policy could fail in a new/perturbed environment that is different from the training environment, due to the presence of dynamic variations. For controlling systems with continuous state and action spaces, we propose an add-on approach to robustifying a pre-trained RL policy by augmenting it with an ${\mathcal {L}_{1}}$ adaptive controller (${\mathcal {L}_{1}}$AC). Leveraging the capability of an ${\mathcal {L}_{1}}$AC for fast estimation and active compensation of dynamic variations, the proposed approach can improve the robustness of an RL policy which is trained either in a simulator or in the real world without consideration of a broad class of dynamic variations. Numerical and real-world experiments empirically demonstrate the efficacy of the proposed approach in robustifying RL policies trained using both model-free and model-based methods.
KW - machine learning for robot control
KW - Reinforcement learning
KW - robot safety
KW - robust/adaptive control
UR - http://www.scopus.com/inward/record.url?scp=85129171219&partnerID=8YFLogxK
U2 - 10.1109/LRA.2022.3169309
DO - 10.1109/LRA.2022.3169309
M3 - Article
AN - SCOPUS:85129171219
SN - 2377-3766
VL - 7
SP - 6574
EP - 6581
JO - IEEE Robotics and Automation Letters
JF - IEEE Robotics and Automation Letters
IS - 3
ER -