Deep neural networks have revolutionized multiple fields within computer science. It is important to have a comprehensive understanding of the memory requirements and performance of deep networks on low-resource systems. While there have been efforts to this end, the effects of severe memory limits and heavy swapping are understudied. We have profiled multiple deep networks under varying memory restrictions and on different hardware. Using this data, we develop two modeling approaches to predict the execution time of a network based on a description of its layers and the available memory. The first modeling approach is based on engineering predictive features through a theoretical analysis of the computations required to execute a layer. Th...
Deep learning models have replaced conventional methods for machine learning tasks. Efficient infere...
Deep neural networks (DNNs) are becoming a key enabling technique for many application domains. Howe...
Hardware accelerators for neural network inference can exploit common data properties for performanc...
Deep learning is attracting interest across a variety of domains, including natural language process...
Abstract. We present an estimation methodology, accurately predicting the execution time for a given...
This thesis proposes a convolutional long short-term memory neural network model for predicting limi...
Large data transfers are getting more critical with the increasing volume of data in scientific comp...
Performance models for storage devices are an important part of simulations of large-scale computing...
The final publication is available at ACM via http://dx.doi.org/10.1145/3229607.3229613Recent trends...
We devise a performance model for GPU training of Deep Learning Recommendation Models (DLRM), whose ...
Neural networks have been widely applied to various research and production fields. However, most re...
Accurately modeling and predicting performance for large-scale applications becomes increasingly dif...
There are still many problems that need to be solved with Internet of Things (IoT) technology, one o...
© 2021 IEEE.To meet surging demands for deep learning inference services, many cloud computing vendo...
International audienceMachine learning is one of the most cutting edge methods in computer vision. C...
Deep learning models have replaced conventional methods for machine learning tasks. Efficient infere...
Deep neural networks (DNNs) are becoming a key enabling technique for many application domains. Howe...
Hardware accelerators for neural network inference can exploit common data properties for performanc...
Deep learning is attracting interest across a variety of domains, including natural language process...
Abstract. We present an estimation methodology, accurately predicting the execution time for a given...
This thesis proposes a convolutional long short-term memory neural network model for predicting limi...
Large data transfers are getting more critical with the increasing volume of data in scientific comp...
Performance models for storage devices are an important part of simulations of large-scale computing...
The final publication is available at ACM via http://dx.doi.org/10.1145/3229607.3229613Recent trends...
We devise a performance model for GPU training of Deep Learning Recommendation Models (DLRM), whose ...
Neural networks have been widely applied to various research and production fields. However, most re...
Accurately modeling and predicting performance for large-scale applications becomes increasingly dif...
There are still many problems that need to be solved with Internet of Things (IoT) technology, one o...
© 2021 IEEE.To meet surging demands for deep learning inference services, many cloud computing vendo...
International audienceMachine learning is one of the most cutting edge methods in computer vision. C...
Deep learning models have replaced conventional methods for machine learning tasks. Efficient infere...
Deep neural networks (DNNs) are becoming a key enabling technique for many application domains. Howe...
Hardware accelerators for neural network inference can exploit common data properties for performanc...