首页> 外文会议>International Conference on Artificial Neural Networks >Simplified Computation and Interpretation of Fisher Matrices in Incremental Learning with Deep Neural Networks
【24h】

Simplified Computation and Interpretation of Fisher Matrices in Incremental Learning with Deep Neural Networks

机译:深度神经网络增量学习中Fisher矩阵的简化计算和解释

获取原文

摘要

Import recent advances in the domain of incremental or continual learning with DNNs, such as Elastic Weight Consolidation (EWC) or Incremental Moment Matching (IMM) rely on a quantity termed the Fisher information matrix (FIM). While the results obtained in this way are very promising, the use of the FIM relies on the assumptions that (a) the FIM can be approximated by its diagonal, and (b) that FIM diagonal entries are related to the variance of a DNN parameter in the context of Bayesian neural networks. In addition, the FIM is notoriously difficult to compute in automatic differentiation (AD) systems frameworks like TensorFlow, and existing implementations require an excessive use of memory due to this problem. We present the Matrix of SQuares (MaSQ), computed similarly as the FIM, but whose use in EWC-like algorithms follows directly from the calculus of derivatives and requires no additional assumptions. Additionally, MaSQ computation in AD frameworks is much simpler and more memory-efficient FIM computation. When using MaSQ together with EWC we show superior or equal performance to FIM/EWC on a variety of benchmark tasks.
机译:使用DNN引入增量或连续学习领域的最新进展,例如弹性权重合并(EWC)或增量矩匹配(IMM)都依赖称为Fisher信息矩阵(FIM)的数量。尽管以这种方式获得的结果很有希望,但是FIM的使用依赖于以下假设:(a)FIM可以通过其对角线近似,并且(b)FIM对角线条目与DNN参数的方差有关在贝叶斯神经网络中。此外,众所周知,FIM在诸如TensorFlow之类的自动差分(AD)系统框架中难以计算,并且由于此问题,现有的实现方式需要过多使用内存。我们介绍了SQUAES矩阵(MaSQ),其计算方式与FIM相似,但其在类似EWC的算法中的使用直接来自于导数的演算,不需要任何其他假设。此外,AD框架中的MaSQ计算是更简单且内存效率更高的FIM计算。当将MaSQ与EWC一起使用时,在各种基准测试任务上,我们表现出优于FIM / EWC的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号