DGCN: Dynamic Graph Convolutional Network for Efficient Multi-Person Pose Estimation

Authors

  • Zhongwei Qiu University of Science and Technology Beijing
  • Kai Qiu Microsoft Research Asia
  • Jianlong Fu Microsoft Research Asia
  • Dongmei Fu University of Science and Technology Beijing

DOI:

https://doi.org/10.1609/aaai.v34i07.6867

Abstract

Multi-person pose estimation aims to detect human keypoints from images with multiple persons. Bottom-up methods for multi-person pose estimation have attracted extensive attention, owing to the good balance between efficiency and accuracy. Recent bottom-up methods usually follow the principle of keypoints localization and grouping, where relations between keypoints are the keys to group keypoints. These relations spontaneously construct a graph of keypoints, where the edges represent the relations between two nodes (i.e., keypoints). Existing bottom-up methods mainly define relations by empirically picking out edges from this graph, while omitting edges that may contain useful semantic relations. In this paper, we propose a novel Dynamic Graph Convolutional Module (DGCM) to model rich relations in the keypoints graph. Specifically, we take into account all relations (all edges of the graph) and construct dynamic graphs to tolerate large variations of human pose. The DGCM is quite lightweight, which allows it to be stacked like a pyramid architecture and learn structural relations from multi-level features. Our network with single DGCM based on ResNet-50 achieves relative gains of 3.2% and 4.8% over state-of-the-art bottom-up methods on COCO keypoints and MPII dataset, respectively.

Downloads

Published

2020-04-03

How to Cite

Qiu, Z., Qiu, K., Fu, J., & Fu, D. (2020). DGCN: Dynamic Graph Convolutional Network for Efficient Multi-Person Pose Estimation. Proceedings of the AAAI Conference on Artificial Intelligence, 34(07), 11924-11931. https://doi.org/10.1609/aaai.v34i07.6867

Issue

Section

AAAI Technical Track: Vision