HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Exponential Moving Average Normalization for Self-supervised and Semi-supervised Learning

Zhaowei Cai Avinash Ravichandran Subhransu Maji Charless Fowlkes Zhuowen Tu Stefano Soatto

Exponential Moving Average Normalization for Self-supervised and Semi-supervised Learning

Abstract

We present a plug-in replacement for batch normalization (BN) called exponential moving average normalization (EMAN), which improves the performance of existing student-teacher based self- and semi-supervised learning techniques. Unlike the standard BN, where the statistics are computed within each batch, EMAN, used in the teacher, updates its statistics by exponential moving average from the BN statistics of the student. This design reduces the intrinsic cross-sample dependency of BN and enhances the generalization of the teacher. EMAN improves strong baselines for self-supervised learning by 4-6/1-2 points and semi-supervised learning by about 7/2 points, when 1%/10% supervised labels are available on ImageNet. These improvements are consistent across methods, network architectures, training duration, and datasets, demonstrating the general effectiveness of this technique. The code is available at https://github.com/amazon-research/exponential-moving-average-normalization.

Code Repositories

Benchmarks

BenchmarkMethodologyMetrics
semi-supervised-image-classification-on-1FixMatch-EMAN
Top 1 Accuracy: 63%
semi-supervised-image-classification-on-16FixMatch w/ EMAN (ResNet-50)
ImageNet Top-1 Accuracy: 43.6%
semi-supervised-image-classification-on-2FixMatch-EMAN
Top 1 Accuracy: 74%

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp