Command Palette
Search for a command to run...
HumanBench: Towards General Human-centric Perception with Projector Assisted Pretraining
Shixiang Tang; Cheng Chen; Qingsong Xie; Meilin Chen; Yizhou Wang; Yuanzheng Ci; Lei Bai; Feng Zhu; Haiyang Yang; Li Yi; Rui Zhao; Wanli Ouyang

Abstract
Human-centric perceptions include a variety of vision tasks, which have widespread industrial applications, including surveillance, autonomous driving, and the metaverse. It is desirable to have a general pretrain model for versatile human-centric downstream tasks. This paper forges ahead along this path from the aspects of both benchmark and pretraining methods. Specifically, we propose a \textbf{HumanBench} based on existing datasets to comprehensively evaluate on the common ground the generalization abilities of different pretraining methods on 19 datasets from 6 diverse downstream tasks, including person ReID, pose estimation, human parsing, pedestrian attribute recognition, pedestrian detection, and crowd counting. To learn both coarse-grained and fine-grained knowledge in human bodies, we further propose a \textbf{P}rojector \textbf{A}ssis\textbf{T}ed \textbf{H}ierarchical pretraining method (\textbf{PATH}) to learn diverse knowledge at different granularity levels. Comprehensive evaluations on HumanBench show that our PATH achieves new state-of-the-art results on 17 downstream datasets and on-par results on the other 2 datasets. The code will be publicly at \href{https://github.com/OpenGVLab/HumanBench}{https://github.com/OpenGVLab/HumanBench}.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| pedestrian-attribute-recognition-on-pa-100k | PATH (Partial FT) | Accuracy: 90.8 |
| pose-estimation-on-coco | PATH (Partial FT) | AP: 77.1 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.