HyperAIHyperAI

Command Palette

Search for a command to run...

4 months ago

Architectural Complexity Measures of Recurrent Neural Networks

Saizheng Zhang; Yuhuai Wu; Tong Che; Zhouhan Lin; Roland Memisevic; Ruslan Salakhutdinov; Yoshua Bengio

Architectural Complexity Measures of Recurrent Neural Networks

Abstract

In this paper, we systematically analyze the connecting architectures of recurrent neural networks (RNNs). Our main contribution is twofold: first, we present a rigorous graph-theoretic framework describing the connecting architectures of RNNs in general. Second, we propose three architecture complexity measures of RNNs: (a) the recurrent depth, which captures the RNN's over-time nonlinear complexity, (b) the feedforward depth, which captures the local input-output nonlinearity (similar to the "depth" in feedforward neural networks (FNNs)), and (c) the recurrent skip coefficient which captures how rapidly the information propagates over time. We rigorously prove each measure's existence and computability. Our experimental results show that RNNs might benefit from larger recurrent depth and feedforward depth. We further demonstrate that increasing recurrent skip coefficient offers performance boosts on long term dependency problems.

Benchmarks

BenchmarkMethodologyMetrics
language-modelling-on-text8td-LSTM (Zhang et al., 2016)
Bit per Character (BPC): 1.63
language-modelling-on-text8td-LSTM-large
Bit per Character (BPC): 1.49

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Architectural Complexity Measures of Recurrent Neural Networks | Papers | HyperAI