Neural networks have gained widespread use in many machine learning tasks due to their state-of-the-art performance. However, the cost of this progress lies in the ever-increasing sizes and computational demands of the resulting models. As such, the neural network compression, the process of reducing the size, power consumption, or any other cost of interest of the model, has become an important practical step when deploying the trained models to perform inference tasks. In this dissertation, we explore a particular compression mechanism --- the low-rank decomposition --- and its extensions for the purposes of neural network compression. We study important aspects of the low-rank compression: how to select the decomposition ranks across the...
While deep neural networks are a highly successful model class, their large memory footprint puts co...
In this paper, we study the compression of a target two-layer neural network with N nodes into a com...
[EN] Artificial intelligence algorithms have experienced a dramatic improvement in the last decade ...
Neural networks have gained widespread use in many machine learning tasks due to their state-of-the-...
Recently, the deep neural network (DNN) has become one of the most advanced and powerful methods use...
Low-rankness plays an important role in traditional machine learning, but is not so popular in deep ...
© 2017 IEEE. Deep compression refers to removing the redundancy of parameters and feature maps for d...
Low-rank compression is an important model compression strategy for obtaining compact neural network...
A common technique for compressing a neural network is to compute the k-rank ℓ2 approximation Ak of ...
Compressing neural networks is a key step when deploying models for real-time or embedded applicatio...
Compression of a neural network can help in speeding up both the training and the inference of the n...
In recent years, neural networks have grown in popularity, mostly thanks to the advances in the fiel...
In the last years, deep neural networks have revolutionized machine learning tasks. However, the des...
Lossy gradient compression has become a practical tool to overcome the communication bottleneck in c...
While deep neural networks are a highly successful model class, their large memory footprint puts co...
While deep neural networks are a highly successful model class, their large memory footprint puts co...
In this paper, we study the compression of a target two-layer neural network with N nodes into a com...
[EN] Artificial intelligence algorithms have experienced a dramatic improvement in the last decade ...
Neural networks have gained widespread use in many machine learning tasks due to their state-of-the-...
Recently, the deep neural network (DNN) has become one of the most advanced and powerful methods use...
Low-rankness plays an important role in traditional machine learning, but is not so popular in deep ...
© 2017 IEEE. Deep compression refers to removing the redundancy of parameters and feature maps for d...
Low-rank compression is an important model compression strategy for obtaining compact neural network...
A common technique for compressing a neural network is to compute the k-rank ℓ2 approximation Ak of ...
Compressing neural networks is a key step when deploying models for real-time or embedded applicatio...
Compression of a neural network can help in speeding up both the training and the inference of the n...
In recent years, neural networks have grown in popularity, mostly thanks to the advances in the fiel...
In the last years, deep neural networks have revolutionized machine learning tasks. However, the des...
Lossy gradient compression has become a practical tool to overcome the communication bottleneck in c...
While deep neural networks are a highly successful model class, their large memory footprint puts co...
While deep neural networks are a highly successful model class, their large memory footprint puts co...
In this paper, we study the compression of a target two-layer neural network with N nodes into a com...
[EN] Artificial intelligence algorithms have experienced a dramatic improvement in the last decade ...