Efficient Video Pose Estimation via Neural Architecture Search

Human pose estimation has achieved significant progress in recent years. However, most of the recent methods focus on improving accuracy using complicated models and ignoring real-time efficiency. To achieve a better trade-off between accuracy and efficiency, we propose a novel neural architecture search (NAS) method, termed ViP-NAS, to search networks in both spatial and temporal levels for fast online video pose estimation.

Data and Resources

Cite this as

Lumin Xu, Yingda Guan, Sheng Jin, Wentao Liu, Chen Qian, Ping Luo, Wanli Ouyang, Xiaogang Wang (2025). Dataset: Efficient Video Pose Estimation via Neural Architecture Search. https://doi.org/10.57702/jz11p5eg

DOI retrieved: January 2, 2025

Additional Info

Field Value
Created January 2, 2025
Last update January 2, 2025
Defined In https://doi.org/10.48550/arXiv.2105.10154
Author Lumin Xu
More Authors
Yingda Guan
Sheng Jin
Wentao Liu
Chen Qian
Ping Luo
Wanli Ouyang
Xiaogang Wang
Homepage https://arxiv.org/abs/2006.11145