论文标题
ADG置置:实际人类姿势估计的自动化数据集生成
ADG-Pose: Automated Dataset Generation for Real-World Human Pose Estimation
论文作者
论文摘要
计算机视觉上的最新进展使使用神经网络以了解人类姿势的应用的突出性有所提高。但是,尽管最新数据集的准确性一直在稳步提高,但这些数据集通常无法解决现实世界应用程序中所面临的挑战。这些挑战是与远离镜头的人们,人群中的人们和被遮挡的人打交道。结果,许多现实世界的应用程序已经培训了无法反映部署中存在的数据的数据,从而导致表现不佳。本文介绍了ADG-POSE,这是一种自动生成数据集以进行现实世界姿势估计的方法。可以定制这些数据集以确定人距离,拥挤和遮挡分布。接受我们方法训练的模型能够在存在这些挑战的情况下执行,而这些挑战在其他数据集中训练失败的模型。使用基于现实骨架的动作识别的ADG置端,端到端的精度,在中等距离和遮挡水平的场景上增加了20%,并且在其他模型表现不得比随机性更好的遥远场景上增加了4倍。
Recent advancements in computer vision have seen a rise in the prominence of applications using neural networks to understand human poses. However, while accuracy has been steadily increasing on State-of-the-Art datasets, these datasets often do not address the challenges seen in real-world applications. These challenges are dealing with people distant from the camera, people in crowds, and heavily occluded people. As a result, many real-world applications have trained on data that does not reflect the data present in deployment, leading to significant underperformance. This article presents ADG-Pose, a method for automatically generating datasets for real-world human pose estimation. These datasets can be customized to determine person distances, crowdedness, and occlusion distributions. Models trained with our method are able to perform in the presence of these challenges where those trained on other datasets fail. Using ADG-Pose, end-to-end accuracy for real-world skeleton-based action recognition sees a 20% increase on scenes with moderate distance and occlusion levels, and a 4X increase on distant scenes where other models failed to perform better than random.