Modern deep neural networks must demonstrate state-of-the-art accuracy while exhibiting low latency and energy consumption. As such, neural architecture search (NAS) algorithms take these two constraints into account when generating a new architecture. However, efficiency metrics such as latency are typically hardware dependent requiring the NAS algorithm to either measure or predict the architecture latency. Measuring the latency of every evaluated architecture adds a significant amount of time to the NAS process. Here we propose Microprocessor A Priori for Latency Estimation MAPLE that does not rely on transfer learning or domain adaptation but instead generalizes to new hardware by incorporating a prior hardware characteristics during tr...
Certain automatic designs of neural networks not only minimize prediction error but also shrink or p...
With more powerful yet efficient embedded devices and accelerators being available for Deep Neural N...
CNN design and deployment on embedded edge-processing systems is an error-prone and effort-hungry pr...
Deep neural network (DNN) latency characterization is a time-consuming process and adds significant ...
The ability to accurately predict deep neural network (DNN) inference performance metrics, such as l...
The massive use of artificial neural networks (ANNs), increasingly popular in many areas of scientif...
While providing the same functionality, the various Deep Learning software frameworks available thes...
With the surge of inexpensive computational and memory resources, neural networks (NNs) have experie...
Field-programmable gate array (FPGA) based accelerators are being widely used for acceleration of co...
Neural architecture search (NAS) is an emerging paradigm to automate the design of top-performing de...
DNNs have been finding a growing number of applications including image classification, speech recog...
Deep Neural Networks (DNNs) are extremely computationally demanding, which presents a large barrier ...
Mixed-precision quantization, where a deep neural network's layers are quantized to different precis...
Large Deep Neural Networks (DNNs) are the backbone of today's artificial intelligence due to their a...
Deep neural networks (DNNs) are a vital tool in pattern recognition and Machine Learning (ML) – solv...
Certain automatic designs of neural networks not only minimize prediction error but also shrink or p...
With more powerful yet efficient embedded devices and accelerators being available for Deep Neural N...
CNN design and deployment on embedded edge-processing systems is an error-prone and effort-hungry pr...
Deep neural network (DNN) latency characterization is a time-consuming process and adds significant ...
The ability to accurately predict deep neural network (DNN) inference performance metrics, such as l...
The massive use of artificial neural networks (ANNs), increasingly popular in many areas of scientif...
While providing the same functionality, the various Deep Learning software frameworks available thes...
With the surge of inexpensive computational and memory resources, neural networks (NNs) have experie...
Field-programmable gate array (FPGA) based accelerators are being widely used for acceleration of co...
Neural architecture search (NAS) is an emerging paradigm to automate the design of top-performing de...
DNNs have been finding a growing number of applications including image classification, speech recog...
Deep Neural Networks (DNNs) are extremely computationally demanding, which presents a large barrier ...
Mixed-precision quantization, where a deep neural network's layers are quantized to different precis...
Large Deep Neural Networks (DNNs) are the backbone of today's artificial intelligence due to their a...
Deep neural networks (DNNs) are a vital tool in pattern recognition and Machine Learning (ML) – solv...
Certain automatic designs of neural networks not only minimize prediction error but also shrink or p...
With more powerful yet efficient embedded devices and accelerators being available for Deep Neural N...
CNN design and deployment on embedded edge-processing systems is an error-prone and effort-hungry pr...