首页> 外文OA文献 >A real-time HIL control system on rotary inverted pendulum hardware platform based on double deep Q-network
【2h】

A real-time HIL control system on rotary inverted pendulum hardware platform based on double deep Q-network

机译:基于双深Q网的旋转倒立摆硬件平台实时HIL控制系统

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

For real applications, rotary inverted pendulum systems have been known as the basic model in nonlinear control systems. If researchers have no deep understanding of control, it is difficult to control a rotary inverted pendulum platform using classic control engineering models, as shown in section 2.1. Therefore, without classic control theory, this paper controls the platform by training and testing reinforcement learning algorithm. Many recent achievements in reinforcement learning (RL) have become possible, but there is a lack of research to quickly test high-frequency RL algorithms using real hardware environment. In this paper, we propose a real-time Hardware-in-the-loop (HIL) control system to train and test the deep reinforcement learning algorithm from simulation to real hardware implementation. The Double Deep Q-Network (DDQN) with prioritized experience replay reinforcement learning algorithm, without a deep understanding of classical control engineering, is used to implement the agent. For the real experiment, to swing up the rotary inverted pendulum and make the pendulum smoothly move, we define 21 actions to swing up and balance the pendulum. Comparing Deep Q-Network (DQN), the DDQN with prioritized experience replay algorithm removes the overestimate of Q value and decreases the training time. Finally, this paper shows the experiment results with comparisons of classic control theory and different reinforcement learning algorithms.
机译:对于真实应用,旋转倒立摆系统已被称为非线性控制系统中的基本模型。如果研究人员没有深入了解控制,则难以使用经典控制工程模型来控制旋转倒立摆平台,如第2.1节所示。因此,没有经典控制理论,本文通过培训和测试强化学习算法控制平台。最近近期加强学习(RL)的成就已经成为可能,但缺乏研究使用真实的硬件环境快速测试高频RL算法。在本文中,我们提出了一种实时硬件循环(HIL)控制系统,用于从模拟到实际硬件实现中培训和测试深度加强学习算法。双层Q-Network(DDQN)具有优先考虑的体验重放增强学习算法,无需深入了解古典控制工程,用于实现代理。对于真实的实验,为了使旋转倒立摆动摆动并使摆锤平滑地移动,我们定义了21个动作来摆动并平衡摆锤。比较Deep Q-Network(DQN),具有优先考虑体验的DDQN重放算法消除了Q值的高估并降低了训练时间。最后,本文显示了经典控制理论与不同增强学习算法的比较实验结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号