【24h】

Visual Saliency Estimation through Manifold Learning

机译:通过流形学习进行视觉显着性估计

获取原文

摘要

Saliency detection has been a desirable way for robotic vision to find the most noticeable objects in a scene. In this paper, a robust manifold based saliency estimation method has been developed to help capture the most salient objects in front of robotic eyes, namely cameras. In the proposed approach, an image is considered as a manifold of visual signals (stimuli) spreading over a connected grid, and local visual stimuli are compared against the global image variation to model the visual saliency. With this model, manifold learning is then applied to minimize the local variation while keeping the global contrast, and turns the RGB image into a multi channel image. After the projection through manifold learning, histogram based contrast is then computed for saliency modeling of all channels of the projected images, and mutual information is introduced to evaluate each single channel saliency map against prior knowledge to provide cues for the fusion of multiple channels. In the last step, the fusion procedure combines all single channel saliency maps according to their mutual information score, and generates the final saliency map. In our experiment, the proposed method is evaluated using one of the largest publicly available image datasets. The experimental results validated that our algorithm consistently outperforms the state of the art unsupervised saliency detection methods, yielding higher precision and better recall rates. Furthermore, the proposed method is tested on a video where a moving camera is trying to catch up with the walking person a salient object in the video sequence. Our experimental results demonstrated that the proposed approach can successful accomplish this task, revealing its potential use for similar robotic applications.
机译:显着性检测一直是机器人视觉在场景中找到最引人注目的对象的一种理想方式。在本文中,已经开发了一种基于鲁棒性的显着性估计方法,以帮助捕获机器人眼前最显着的物体,即相机。在提出的方法中,图像被视为传播在连接的网格上的视觉信号(刺激)的流形,并且将局部视觉刺激与全局图像变化进行比较以对视觉显着性进行建模。使用此模型,然后应用流形学习以在保持全局对比度的同时最小化局部变化,并将RGB图像转换为多通道图像。在通过流水线学习进行投影之后,随后将基于直方图的对比度进行计算,以对投影图像的所有通道进行显着性建模,并引入互信息以根据先验知识评估每个单个通道的显着性图,从而为融合多个通道提供线索。在最后一步中,融合过程根据所有单通道显着性图的相互信息得分将它们组合在一起,并生成最终的显着性图。在我们的实验中,使用最大的公开可用图像数据集之一对提出的方法进行了评估。实验结果验证了我们的算法始终优于现有的无监督显着性检测方法,从而产生了更高的精度和更好的召回率。此外,在视频上测试了所提出的方法,其中运动的摄像头正试图赶上步行者视频序列中的一个显着物体。我们的实验结果表明,所提出的方法可以成功完成此任务,从而揭示了其在类似机器人应用中的潜在用途。

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号