首页> 外文会议>2018 International Conference on Cloud Computing, Big Data and Blockchain >Fully Binarized Convolutional Neural Network for Accelerating Edge Vision Computing
【24h】

Fully Binarized Convolutional Neural Network for Accelerating Edge Vision Computing

机译:完全二值化卷积神经网络,用于加速边缘视觉计算

获取原文
获取原文并翻译 | 示例

摘要

We propose BitFlow-Net, a method to simplify binary CNN model to almost no floating-point multiplication at inference time. Recently, a lot of variations of binarized networks were proposed trying to achieve high accuracy while replacing resource-consuming floating-point multiplication with bit operation. These methods usually require scaling factor and BatchNorm to achieve comparable accuracy as their full-precision counterparts. However, data flow have to be frequently converted between floating-point data and bit data due to the multiplication with scaling factor and in BatchNorm. Such conversion will cost extra resources and time when implemented on edge hardware. Motivated by that, we further explore and reveal some basic attributes of BNN based on previous works and propose a new method to simplify binary network. As a result, our model could inference with most of its data flow remains bit flow. Such a network architecture will greatly reduce the design complexity when implemented on ASIC or FPGA. Our method performs no accuracy degradation on ImageNet compared to state-of-the-art BNN models but without extra floating-point multiplications.
机译:我们提出了BitFlow-Net,这是一种将二进制CNN模型简化为在推理时几乎没有浮点乘法的方法。近来,提出了许多二值化网络的变体,试图在用位运算代替消耗资源的浮点乘法的同时实现高精度。这些方法通常需要比例因子和BatchNorm才能获得与全精度同类产品相当的精度。但是,由于与比例因子和BatchNorm相乘,因此必须经常在浮点数据和位数据之间转换数据流。当在边缘硬件上实施时,这种转换将花费额外的资源和时间。为此,我们在先前工作的基础上进一步探索并揭示了BNN的一些基本属性,并提出了一种简化二进制网络的新方法。结果,我们的模型可以推断出其大部分数据流仍为位流。当在ASIC或FPGA上实现时,这样的网络体系结构将大大降低设计复杂性。与最新的BNN模型相比,我们的方法不会对ImageNet造成任何精度下降,但无需额外的浮点乘法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号