【24h】

HCVRD: A Benchmark for Large-Scale Human-Centered Visual Relationship Detection

机译:HCVRD:大规模人以人为本的视觉关系检测的基准

获取原文

摘要

Visual relationship detection aims to capture interactions between pairs of objects in images. Relationships between objects and humans represent a particularly important subset of this problem, with implications for challenges such as understanding human behavior, and identifying affordances, amongst others. In addressing this problem we first construct a large-scale human-centric visual relationship detection dataset (HCVRD), which provides many more types of relationship annotations (nearly 10K categories) than the previous released datasets. This large label space better reflects the reality of human-object interactions, but gives rise to a long-tail distribution problem, which in turn demands a zero-shot approach to labels appearing only in the test set. This is the first time this issue has been addressed. We propose a webly-supervised approach to these problems and demonstrate that the proposed model provides a strong baseline on our HCVRD dataset.
机译:视觉关系检测旨在捕获图像中对象对之间的相互作用。 物体与人类之间的关系代表了这个问题的特别重要的子集,对诸如了解人类行为的挑战,以及识别带来的挑战,以及别人的影响。 在解决此问题时,我们首先构造一个大规模的人类视觉关系检测数据集(HCVRD),它提供了比以前发布的数据集更多类型的关系注释(近10k类别)。 这个大型标签空间更好地反映了人对象交互的现实,而是产生了长尾分布问题,这反过来又需要一个仅在测试集中出现的标签的零点方法。 这是第一次解决了这个问题。 我们为这些问题提出了一种令人柔和的监督方法,并证明了拟议的模型在我们的HCVRD数据集中提供了强大的基准。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号