HyperAIHyperAI

Command Palette

Search for a command to run...

4 months ago

Tensorizing Neural Networks

Alexander Novikov; Dmitry Podoprikhin; Anton Osokin; Dmitry Vetrov

Tensorizing Neural Networks

Abstract

Deep neural networks currently demonstrate state-of-the-art performance in several domains. At the same time, models of this class are very demanding in terms of computational resources. In particular, a large amount of memory is required by commonly used fully-connected layers, making it hard to use the models on low-end devices and stopping the further increase of the model size. In this paper we convert the dense weight matrices of the fully-connected layers to the Tensor Train format such that the number of parameters is reduced by a huge factor and at the same time the expressive power of the layer is preserved. In particular, for the Very Deep VGG networks we report the compression factor of the dense weight matrix of a fully-connected layer up to 200000 times leading to the compression factor of the whole network up to 7 times.

Code Repositories

timgaripov/TensorNet-TF
tf
Mentioned in GitHub
Bihaqo/TensorNet
Official
tf
Mentioned in GitHub
Gyiming/MobileSLAM
tf
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
image-classification-on-mnistPerceptron with a tensor train layer
Accuracy: 98.2
Percentage error: 1.8

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp