HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Efficient generative adversarial networks using linear additive-attention Transformers

Emilio Morales-Juarez Gibran Fuentes-Pineda

Efficient generative adversarial networks using linear additive-attention Transformers

Abstract

Although the capacity of deep generative models for image generation, such as Diffusion Models (DMs) and Generative Adversarial Networks (GANs), has dramatically improved in recent years, much of their success can be attributed to computationally expensive architectures. This has limited their adoption and use to research laboratories and companies with large resources, while significantly raising the carbon footprint for training, fine-tuning, and inference. In this work, we present a novel GAN architecture which we call LadaGAN. This architecture is based on a linear attention Transformer block named Ladaformer. The main component of this block is a linear additive-attention mechanism that computes a single attention vector per head instead of the quadratic dot-product attention. We employ Ladaformer in both the generator and discriminator, which reduces the computational complexity and overcomes the training instabilities often associated with Transformer GANs. LadaGAN consistently outperforms existing convolutional and Transformer GANs on benchmark datasets at different resolutions while being significantly more efficient. Moreover, LadaGAN shows competitive performance compared to state-of-the-art multi-step generative models (e.g. DMs) using orders of magnitude less computational resources.

Code Repositories

milmor/LadaGAN-pytorch
Official
pytorch
Mentioned in GitHub
milmor/ladagan
Official
tf
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
image-generation-on-celeba-64x64LadaGAN
FID: 1.81
image-generation-on-ffhq-128-x-128LadaGAN
FID: 4.48
image-generation-on-lsun-bedroom-128-x-128LadaGAN
FID: 4.90

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp