首页> 外文会议>International Conference on Artificial Neural Networks >Simplified Computation and Interpretation of Fisher Matrices in Incremental Learning with Deep Neural Networks
【24h】

Simplified Computation and Interpretation of Fisher Matrices in Incremental Learning with Deep Neural Networks

机译:具有深神经网络增量学习中Fisher矩阵的简化计算与解释

获取原文

摘要

Import recent advances in the domain of incremental or continual learning with DNNs, such as Elastic Weight Consolidation (EWC) or Incremental Moment Matching (IMM) rely on a quantity termed the Fisher information matrix (FIM). While the results obtained in this way are very promising, the use of the FIM relies on the assumptions that (a) the FIM can be approximated by its diagonal, and (b) that FIM diagonal entries are related to the variance of a DNN parameter in the context of Bayesian neural networks. In addition, the FIM is notoriously difficult to compute in automatic differentiation (AD) systems frameworks like TensorFlow, and existing implementations require an excessive use of memory due to this problem. We present the Matrix of SQuares (MaSQ), computed similarly as the FIM, but whose use in EWC-like algorithms follows directly from the calculus of derivatives and requires no additional assumptions. Additionally, MaSQ computation in AD frameworks is much simpler and more memory-efficient FIM computation. When using MaSQ together with EWC we show superior or equal performance to FIM/EWC on a variety of benchmark tasks.
机译:使用DNN的增量或持续学习领域的进口近期进步,例如弹性重量整合(EWC)或增量时刻匹配(IMM)依赖于被称为Fisher信息矩阵(FIM)的数量。虽然以这种方式获得的结果非常有前途,但FIM的使用依赖于(a)FIM可以通过其对角线近似的假设,并且(b)FIM对角线条目与DNN参数的方差有关在贝叶斯神经网络的背景下。此外,FIM难以在自动分化(AD)系统框架中计算,如TensorFlow,并且现有的实现导致由于此问题而过度使用内存。我们介绍了正方形(MASQ)的矩阵,类似地计算为FIM,但其在EWC样算法中的用途直接从衍生物的微积分中遵循并且不需要额外的假设。此外,广告框架中的MASQ计算更简单,更高的内存高效计算。使用MASQ与EWC一起使用我们对各种基准任务的FIM / EWC显示出优越或平等的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号