Predicting the performance of Artificial Neural Networks (ANNs) on embedded multi-core platforms is tedious. Concurrent accesses to shared resources are hard to model due to congestion effects on the shared communication medium, which affect the performance of the application. In this paper we present a hybrid modeling environment to enable fast yet accurate timing prediction for fully-connected ANNs deployed on multi-core platforms. The modeling flow is based on the integration of an analytical computation time model with a communication time model which are both calibrated through measurement inside a system level simulation using SystemC. The proposed flow enables the prediction of the end-to-end latency for different mappings of several...
© 2021 IEEE.To meet surging demands for deep learning inference services, many cloud computing vendo...
Deep neural network (DNN) latency characterization is a time-consuming process and adds significant ...
Shared memory multiprocessors require reconfigurable interconnection networks (INs) for scalability...
Predicting the performance of Artificial Neural Networks (ANNs) on embedded multi-core platforms is ...
National audiencePredicting the performance of Artificial NeuralNetworks (ANNs) on embedded multi-co...
International audiencePredicting the performance of Artificial Neural Networks(ANNs) on embedded mul...
Evaluation of performance for complex applications such as Artificial Intelligence (AI) algorithms a...
International audienceWhen deploying Artificial Neural Networks (ANNs) onto multi-core embedded plat...
National audienceEvaluation of performance for complex applicationssuch as Artificial Intelligence (...
Early evaluation of Neural Networks (NN) deployments on multi-core platforms is necessary to find de...
Abstract. We present an estimation methodology, accurately predicting the execution time for a given...
. A performance prediction method is presented for indicating the performance range of MIMD parallel...
Diminishing performance returns and increasing power consumption of single-threaded processors have ...
Deep neural networks have revolutionized multiple fields within computer science. It is important to...
In recent years, machine learning (ML) and, more noticeably, deep learning (DL), have be- come incre...
© 2021 IEEE.To meet surging demands for deep learning inference services, many cloud computing vendo...
Deep neural network (DNN) latency characterization is a time-consuming process and adds significant ...
Shared memory multiprocessors require reconfigurable interconnection networks (INs) for scalability...
Predicting the performance of Artificial Neural Networks (ANNs) on embedded multi-core platforms is ...
National audiencePredicting the performance of Artificial NeuralNetworks (ANNs) on embedded multi-co...
International audiencePredicting the performance of Artificial Neural Networks(ANNs) on embedded mul...
Evaluation of performance for complex applications such as Artificial Intelligence (AI) algorithms a...
International audienceWhen deploying Artificial Neural Networks (ANNs) onto multi-core embedded plat...
National audienceEvaluation of performance for complex applicationssuch as Artificial Intelligence (...
Early evaluation of Neural Networks (NN) deployments on multi-core platforms is necessary to find de...
Abstract. We present an estimation methodology, accurately predicting the execution time for a given...
. A performance prediction method is presented for indicating the performance range of MIMD parallel...
Diminishing performance returns and increasing power consumption of single-threaded processors have ...
Deep neural networks have revolutionized multiple fields within computer science. It is important to...
In recent years, machine learning (ML) and, more noticeably, deep learning (DL), have be- come incre...
© 2021 IEEE.To meet surging demands for deep learning inference services, many cloud computing vendo...
Deep neural network (DNN) latency characterization is a time-consuming process and adds significant ...
Shared memory multiprocessors require reconfigurable interconnection networks (INs) for scalability...