...
首页> 外文期刊>IEEE Transactions on Neural Networks >A Constrained Optimization Approach to Preserving Prior Knowledge During Incremental Training
【24h】

A Constrained Optimization Approach to Preserving Prior Knowledge During Incremental Training

机译:在增量训练中保留先验知识的约束优化方法

获取原文
获取原文并翻译 | 示例
           

摘要

In this paper, a supervised neural network training technique based on constrained optimization is developed for preserving prior knowledge of an input–output mapping during repeated incremental training sessions. The prior knowledge, referred to as long-term memory (LTM), is expressed in the form of equality constraints obtained by means of an algebraic training technique. Incremental training, which may be used to learn new short-term memories (STMs) online, is then formulated as an error minimization problem subject to equality constraints. The solution of this problem is simplified by implementing an adjoined error gradient that circumvents direct substitution and exploits classical backpropagation. A target application is neural network function approximation in adaptive critic designs. For illustrative purposes, constrained training is implemented to update an adaptive critic flight controller, while preserving prior knowledge of an established performance baseline that consists of classical gain-scheduled controllers. It is shown both analytically and numerically that the LTM is accurately preserved while the controller is repeatedly trained over time to assimilate new STMs.
机译:在本文中,开发了一种基于约束优化的有监督的神经网络训练技术,用于在重复的增量训练过程中保留输入输出映射的先验知识。被称为长期记忆(LTM)的先验知识以通过代数训练技术获得的等式约束的形式表示。增量训练可用于在线学习新的短期记忆(STM),然后将其表示为受等式约束的错误最小化问题。通过实现避免直接替换并利用经典反向传播的伴随误差梯度,可以简化此问题的解决方案。目标应用是自适应评论家设计中的神经网络功能逼近。为了说明的目的,实施约束训练以更新自适应评论家飞行控制器,同时保留对由经典增益调度控制器组成的已建立性能基准的先验知识。从分析和数字上都显示,LTM可以准确保存,而随着时间的流逝反复训练控制器以吸收新的STM,LTM可以准确保存。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号