HyperAIHyperAI

Command Palette

Search for a command to run...

4 months ago

Gated Orthogonal Recurrent Units: On Learning to Forget

Li Jing; Caglar Gulcehre; John Peurifoy; Yichen Shen; Max Tegmark; Marin Soljačić; Yoshua Bengio

Gated Orthogonal Recurrent Units: On Learning to Forget

Abstract

We present a novel recurrent neural network (RNN) based model that combines the remembering ability of unitary RNNs with the ability of gated RNNs to effectively forget redundant/irrelevant information in its memory. We achieve this by extending unitary RNNs with a gating mechanism. Our model is able to outperform LSTMs, GRUs and Unitary RNNs on several long-term dependency benchmark tasks. We empirically both show the orthogonal/unitary RNNs lack the ability to forget and also the ability of GORU to simultaneously remember long term dependencies while forgetting irrelevant information. This plays an important role in recurrent neural networks. We provide competitive results along with an analysis of our model on many natural sequential tasks including the bAbI Question Answering, TIMIT speech spectrum prediction, Penn TreeBank, and synthetic tasks that involve long-term dependencies such as algorithmic, parenthesis, denoising and copying tasks.

Code Repositories

jingli9111/GORU-tensorflow
Official
tf
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
question-answering-on-babiGORU
Accuracy (trained on 1k): 60%

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp