This thesis presents a few methods to accelerate the inference of Deep Neural Networks that are large and sparse using GPUs. Deep Neural Networks are now widely used in many applications in various fields, such as computer vision and speech recognition. Deep Neural Networks tend to work more accurately when the model is larger with more layers and neurons, but this makes the model size grow, which causes problems in transferring the data and storing the model in limited fast memory, and it also increases the number of computations, which slows the speed of network inference. The first problem can be solved by using sparse networks with comparable accuracy that contain less weights and thus are smaller in size, and this thesis intends...
Deep neural networks have proven to be particularly effective in visual and audio recognition tasks....
Neural networks get more difficult and longer time to train if the depth become deeper. As deep neur...
Recently, sparse training methods have started to be established as a de facto approach for training...
Convolutional neural networks (CNNs) are often pruned to achieve faster training and inference speed...
In trained deep neural networks, unstructured pruning can reduce redundant weights to lower storage ...
This work is focused on the pruning of some convolutional neural networks (CNNs) and improving their...
Artificial neural networks (ANNs) have emerged as hot topics in the research community. Despite the ...
In recent years, Deep Neural Networks (DNNs) have become an area of high interest due to it's ground...
Machine learning has been widely used in various application domains such as recommendation, compute...
Over the past few years, deep neural networks have been at the center of attention in machine learn...
Hardware accelerators for neural network inference can exploit common data properties for performanc...
High computational complexity and large memory footprint hinder the adoption of convolution neural n...
Doctor of PhilosophyDepartment of Computer ScienceArslan MunirDeep neural networks (DNNs) have gaine...
Accelerating and scaling the training of deep neural networks (DNNs) is critical to keep up with gro...
Machine learning has achieved great success in recent years, especially the deep learning algorithms...
Deep neural networks have proven to be particularly effective in visual and audio recognition tasks....
Neural networks get more difficult and longer time to train if the depth become deeper. As deep neur...
Recently, sparse training methods have started to be established as a de facto approach for training...
Convolutional neural networks (CNNs) are often pruned to achieve faster training and inference speed...
In trained deep neural networks, unstructured pruning can reduce redundant weights to lower storage ...
This work is focused on the pruning of some convolutional neural networks (CNNs) and improving their...
Artificial neural networks (ANNs) have emerged as hot topics in the research community. Despite the ...
In recent years, Deep Neural Networks (DNNs) have become an area of high interest due to it's ground...
Machine learning has been widely used in various application domains such as recommendation, compute...
Over the past few years, deep neural networks have been at the center of attention in machine learn...
Hardware accelerators for neural network inference can exploit common data properties for performanc...
High computational complexity and large memory footprint hinder the adoption of convolution neural n...
Doctor of PhilosophyDepartment of Computer ScienceArslan MunirDeep neural networks (DNNs) have gaine...
Accelerating and scaling the training of deep neural networks (DNNs) is critical to keep up with gro...
Machine learning has achieved great success in recent years, especially the deep learning algorithms...
Deep neural networks have proven to be particularly effective in visual and audio recognition tasks....
Neural networks get more difficult and longer time to train if the depth become deeper. As deep neur...
Recently, sparse training methods have started to be established as a de facto approach for training...