首页> 外文会议>Asia-Pacific Signal and Information Processing Association Annual Summit and Conference >Multi-View and Multi-Modal Action Recognition with Learned Fusion
【24h】

Multi-View and Multi-Modal Action Recognition with Learned Fusion

机译:具有学习融合的多视图和多模式动作识别

获取原文

摘要

In this paper, we study multi-modal and multi-view action recognition system based on the deep-learning techniques. We extended the Temporal Segment Network with additional data fusion stage to combine information from different sources. In this research, we use multiple types of information from different modality such as RGB, depth, infrared data to detect predefined human actions. We tested various combinations of these data sources to examine their impact on the final detection accuracy. We designed 3 information fusion methods to generate the final decision. The most interested one is the Learned Fusion Net designed by us. It turns out the Learned Fusion structure has the best results but requires more training.
机译:在本文中,我们研究了基于深度学习技术的多模式多视图动作识别系统。我们扩展了时间分段网络,并增加了数据融合阶段,以合并来自不同来源的信息。在这项研究中,我们使用来自不同模式的多种信息,例如RGB,深度,红外数据来检测预定义的人类行为。我们测试了这些数据源的各种组合,以检查它们对最终检测精度的影响。我们设计了3种信息融合方法来生成最终决策。最感兴趣的是我们设计的Learned Fusion Net。事实证明,Learned Fusion结构具有最佳结果,但需要更多培训。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号