HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Gradient penalty from a maximum margin perspective

Alexia Jolicoeur-Martineau Ioannis Mitliagkas

Gradient penalty from a maximum margin perspective

Abstract

A popular heuristic for improved performance in Generative adversarial networks (GANs) is to use some form of gradient penalty on the discriminator. This gradient penalty was originally motivated by a Wasserstein distance formulation. However, the use of gradient penalty in other GAN formulations is not well motivated. We present a unifying framework of expected margin maximization and show that a wide range of gradient-penalized GANs (e.g., Wasserstein, Standard, Least-Squares, and Hinge GANs) can be derived from this framework. Our results imply that employing gradient penalties induces a large-margin classifier (thus, a large-margin discriminator in GANs). We describe how expected margin maximization helps reduce vanishing gradients at fake (generated) samples, a known problem in GANs. From this framework, we derive a new $L^\infty$ gradient norm penalty with Hinge loss which generally produces equally good (or better) generated output in GANs than $L^2$-norm penalties (based on the Fréchet Inception Distance).

Code Repositories

lucidrains/stylegan2-pytorch
pytorch
Mentioned in GitHub
AlexiaJM/MaximumMarginGANs
Official
tf
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
image-generation-on-cifar-10HingeGAN
FID: 27.12

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp