首页> 外文会议>International Conference on Artificial Neural Networks >A Study on Catastrophic Forgetting in Deep LSTM Networks
【24h】

A Study on Catastrophic Forgetting in Deep LSTM Networks

机译:深度LSTM网络中的灾难性遗忘研究

获取原文

摘要

We present a systematic study of Catastrophic Forgetting (CF), i.e., the abrupt loss of previously acquired knowledge, when retraining deep recurrent LSTM networks with new samples. CF has recently received renewed attention in the case of feed-forward DNNs, and this article is the first work that aims to rigorously establish whether deep LSTM networks are afflicted by CF as well, and to what degree. In order to test this fully, training is conducted using a wide variety of high-dimensional image-based sequence classification tasks derived from established visual classification benchmarks (MNIST, Devanagari, Fash-ionMNIST and EMNIST). We find that the CF effect occurs universally, without exception, for deep LSTM-based sequence classifiers, regardless of the construction and provenance of sequences. This leads us to conclude that LSTMs, just like DNNs, are fully affected by CF, and that further research work needs to be conducted in order to determine how to avoid this effect (which is not a goal of this study).
机译:我们提供了对灾难性遗忘(CF)的系统研究,即当使用新样本重新训练深度递归LSTM网络时,先前获得的知识突然丢失。 CF最近在前馈DNN方面受到了新的关注,本文是第一篇旨在严格确定深层LSTM网络是否也受CF影响以及在何种程度上受到影响的工作。为了对此进行全面测试,使用从建立的视觉分类基准(MNIST,Devanagari,Fash-ionMNIST和EMNIST)派生的各种基于高维图像的序列分类任务进行训练。我们发现,对于基于LSTM的深度分类器而言,CF效应无一例外地普遍发生,而与序列的构造和出处无关。这使我们得出结论,就像DNN一样,LSTM完全受CF影响,因此需要进行进一步的研究工作以确定如何避免这种影响(这不是本研究的目的)。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号