Command Palette
Search for a command to run...
A Large-Scale Study on Unsupervised Spatiotemporal Representation Learning
Christoph Feichtenhofer; Haoqi Fan; Bo Xiong; Ross Girshick; Kaiming He

Abstract
We present a large-scale study on unsupervised spatiotemporal representation learning from videos. With a unified perspective on four recent image-based frameworks, we study a simple objective that can easily generalize all these methods to space-time. Our objective encourages temporally-persistent features in the same video, and in spite of its simplicity, it works surprisingly well across: (i) different unsupervised frameworks, (ii) pre-training datasets, (iii) downstream datasets, and (iv) backbone architectures. We draw a series of intriguing observations from this study, e.g., we discover that encouraging long-spanned persistency can be effective even if the timespan is 60 seconds. In addition to state-of-the-art results in multiple benchmarks, we report a few promising cases in which unsupervised pre-training can outperform its supervised counterpart. Code is made available at https://github.com/facebookresearch/SlowFast
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| self-supervised-action-recognition-on-hmdb51 | pBYOL | Frozen: false Pre-Training Dataset: Kinetics400 Top-1 Accuracy: 75.0 |
| self-supervised-action-recognition-on-ucf101 | pBYOL | 3-fold Accuracy: 96.3 Frozen: false Pre-Training Dataset: Kinetics400 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.