Deep Learning (DL) is gaining prominence and is widely used for a plethora of problems. DL models, however, take in the order of days to train. Optimizing hyper-parameters is another factor that adds to the training time. This thesis aims to analyze the training pattern on Convolutional Neural Networks from a systems perspective. We perform a thorough study on the effects of systems resources like DRAM, persistent storage (SSD/HDD space), and GPU on the training time. We explore how one could avoid bottlenecks in the data processing pipeline in the training phase. Our analysis illustrates how GPU utilization can be maximized in the training pipeline by choosing the right combination of two hyper-parameters - batch size and the number of dat...
In recent years, machine learning (ML) and, more noticeably, deep learning (DL), have be- come incre...
Deep Neural Network (DNN) frameworks use distributed training to enable faster time to convergence a...
Memory usage is becoming an increasingly pressing bottleneck in the training process of Deep Neural ...
Deep Learning (DL) is gaining prominence and is widely used for a plethora of problems. DL models, h...
Deep Learning, specifically Deep Neural Networks (DNNs), is stressing storage systems in new...
Deep learning is an emerging workload in the field of HPC. This powerful method of resolution is abl...
There has been a recent emergence of applications from the domain of machine learning, data mining, ...
Machine learning brings opportunities for designing efficient computer systems by potentially identi...
Deep neural network models are commonly used in various real-life applications due to their high pre...
Recent decades have witnessed the breakthrough of deep learning algorithms, which have been widely u...
Deep learning has been widely adopted for different applications of artificial intelligence-speech r...
The lifecycle of a deep learning application consists of five phases: Data collection, Architecture ...
We devise a performance model for GPU training of Deep Learning Recommendation Models (DLRM), whose ...
Deep learning has been a very popular topic in Artificial Intelligent industry these years and can b...
While Deep Neural Networks (DNNs) have recently achieved impressive results on many classification t...
In recent years, machine learning (ML) and, more noticeably, deep learning (DL), have be- come incre...
Deep Neural Network (DNN) frameworks use distributed training to enable faster time to convergence a...
Memory usage is becoming an increasingly pressing bottleneck in the training process of Deep Neural ...
Deep Learning (DL) is gaining prominence and is widely used for a plethora of problems. DL models, h...
Deep Learning, specifically Deep Neural Networks (DNNs), is stressing storage systems in new...
Deep learning is an emerging workload in the field of HPC. This powerful method of resolution is abl...
There has been a recent emergence of applications from the domain of machine learning, data mining, ...
Machine learning brings opportunities for designing efficient computer systems by potentially identi...
Deep neural network models are commonly used in various real-life applications due to their high pre...
Recent decades have witnessed the breakthrough of deep learning algorithms, which have been widely u...
Deep learning has been widely adopted for different applications of artificial intelligence-speech r...
The lifecycle of a deep learning application consists of five phases: Data collection, Architecture ...
We devise a performance model for GPU training of Deep Learning Recommendation Models (DLRM), whose ...
Deep learning has been a very popular topic in Artificial Intelligent industry these years and can b...
While Deep Neural Networks (DNNs) have recently achieved impressive results on many classification t...
In recent years, machine learning (ML) and, more noticeably, deep learning (DL), have be- come incre...
Deep Neural Network (DNN) frameworks use distributed training to enable faster time to convergence a...
Memory usage is becoming an increasingly pressing bottleneck in the training process of Deep Neural ...