论文标题
通过知识转移几次射击对象检测
Few-Shot Object Detection via Knowledge Transfer
论文作者
论文摘要
常规的对象检测方法通常需要大量的训练数据和注释的边界框。如果只有少数培训数据和注释,则对象检测器很容易过度且无法概括。它暴露了对象探测器的实际弱点。另一方面,人类只能使用以前学习的知识来轻松地掌握新的推理规则。在本文中,我们通过知识转移介绍了一些射击对象检测,该检测旨在从一些培训示例中检测对象。我们方法的核心是带有附着的元学习者的典型知识转移。元学习者采用支持集图像,其中包括新类别和基本类别的几个示例,并预测将每个类别表示为向量的原型。然后,原型将每个ROI(利益区域)从查询图像到重塑R-CNN预测器头。为了促进重塑过程,我们预测图形结构下的原型,该原型传播了与新类别相关的基本类别的信息,并明确地指导了代表类别之间相关性的先验知识的指导。 Pascal VOC数据集的大量实验验证了所提出的方法的有效性。
Conventional methods for object detection usually require substantial amounts of training data and annotated bounding boxes. If there are only a few training data and annotations, the object detectors easily overfit and fail to generalize. It exposes the practical weakness of the object detectors. On the other hand, human can easily master new reasoning rules with only a few demonstrations using previously learned knowledge. In this paper, we introduce a few-shot object detection via knowledge transfer, which aims to detect objects from a few training examples. Central to our method is prototypical knowledge transfer with an attached meta-learner. The meta-learner takes support set images that include the few examples of the novel categories and base categories, and predicts prototypes that represent each category as a vector. Then, the prototypes reweight each RoI (Region-of-Interest) feature vector from a query image to remodels R-CNN predictor heads. To facilitate the remodeling process, we predict the prototypes under a graph structure, which propagates information of the correlated base categories to the novel categories with explicit guidance of prior knowledge that represents correlations among categories. Extensive experiments on the PASCAL VOC dataset verifies the effectiveness of the proposed method.