首页> 外文期刊>Optimal Control Applications and Methods >Continuous-time receding-horizon reinforcement learning and its application to path-tracking control of autonomous ground vehicles
【24h】

Continuous-time receding-horizon reinforcement learning and its application to path-tracking control of autonomous ground vehicles

机译:Continuous-time receding-horizon reinforcement learning and its application to path-tracking control of autonomous ground vehicles

获取原文
获取原文并翻译 | 示例
           

摘要

Reinforcement learning (RL) and approximate dynamic programming (ADP) have been recently studied to solve nonlinear optimal control problems (OCPs) of continuous-time (CT) systems. However, online learning efficiency and reliability are two major concerns to be further improved. Motivated by the above issues, in this paper we propose a receding-horizon reinforcement learning (RHRL) algorithm for near-optimal control of CT systems under control constraints. Different from classic RL and ADP, in the proposed approach, the infinite-horizon OCP is decomposed as a series of finite-horizon ones solved with an actor-critic structure according to the receding horizon strategy, which can improve the online learning efficiency and reliability. The unknown dynamics of the system are identified offline using a sparse kernel-based neural network structure whose weights are also updated online in the RHRL framework to improve the control performance. Moreover, the convergence of the modeling error is proven. To verify the effectiveness of our approach, we apply the RHRL algorithm to the autonomous ground vehicle for realizing near-optimal path-tracking control. Compared with CT model predictive control using a nominal model and other model-free tracking controllers such as pure pursuit, heuristic dual programming, and the soft actor-critic algorithm, RHRL performs better in terms of control performance.

著录项

获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号