HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Temporal FiLM: Capturing Long-Range Sequence Dependencies with Feature-Wise Modulations.

{Zayd Enam Sawyer Birnbaum Volodymyr Kuleshov Pang Wei W. Koh Stefano Ermon}

Temporal FiLM: Capturing Long-Range Sequence Dependencies with Feature-Wise Modulations.

Abstract

Learning representations that accurately capture long-range dependencies in sequential inputs --- including text, audio, and genomic data --- is a key problem in deep learning. Feed-forward convolutional models capture only feature interactions within finite receptive fields while recurrent architectures can be slow and difficult to train due to vanishing gradients. Here, we propose Temporal Feature-Wise Linear Modulation (TFiLM) --- a novel architectural component inspired by adaptive batch normalization and its extensions --- that uses a recurrent neural network to alter the activations of a convolutional model. This approach expands the receptive field of convolutional sequence models with minimal computational overhead. Empirically, we find that TFiLM significantly improves the learning speed and accuracy of feed-forward neural networks on a range of generative and discriminative learning tasks, including text classification and audio super-resolution.

Benchmarks

BenchmarkMethodologyMetrics
audio-super-resolution-on-vctk-multi-speaker-1U-Net + TFiLM
Log-Spectral Distance: 1.8

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Temporal FiLM: Capturing Long-Range Sequence Dependencies with Feature-Wise Modulations. | Papers | HyperAI