On-device DNN processing has been common interests in the field of autonomous driving research. For better accuracy, both the number of DNN models and the model-complexity have been increased. To properly respond to this, hardware platforms structured with multicore-based CPUs and DNN accelerators have been released, and the GPU is generally used as an accelerator. When multiple DNN workloads are sporadically requested, the GPU can be easily oversubscribed, thereby leading to an unexpected performance bottleneck. We propose an on-device CPU-GPU co-scheduling framework for multi-DNN execution to remove the performance barrier precluding DNN executions from being bounded by the GPU. Our framework fills up the unused CPU cycles with DNN comput...
Part 2: AIInternational audienceConvolutional neural networks (CNNs) are widely used in vision-based...
Mobile and embedded platforms are increasingly required to efficiently execute computationally deman...
Platforms with multiple Field Programmable Gate Arrays (FPGAs), such as Amazon Web Services (AWS) F1...
Nowadays, Deep learning-based solutions and, in particular, deep neural networks (DNNs) are getting ...
Deep learning-based solutions and, in particular, deep neural networks (DNNs) are at the heart of se...
Our work seeks to improve and adapt computing systems and machine learning (ML) algorithms to match ...
RISC-V is an open-source instruction set and now has been examined as a universal standard to unify ...
DNN inference is increasingly being executed locally on embedded platforms, due to the clear advanta...
Many applications such as autonomous driving and augmented reality, require the concurrent running o...
Deep neural network (DNN) inference is increasingly being executed on mobile and embedded platforms ...
Deep neural networks (DNNs) have emerged as successful solutions for variety of artificial intellige...
Thesis (Master's)--University of Washington, 2018Embedded platforms with integrated graphics process...
A plethora of applications are using machine learning, the operations of which are becoming more com...
GPUs are the workhorse in modern server infrastructure fueling advances in a number of compute-inten...
Deep Neural Network (DNN) inference is increasingly being deployed on edge devices, driven by the ad...
Part 2: AIInternational audienceConvolutional neural networks (CNNs) are widely used in vision-based...
Mobile and embedded platforms are increasingly required to efficiently execute computationally deman...
Platforms with multiple Field Programmable Gate Arrays (FPGAs), such as Amazon Web Services (AWS) F1...
Nowadays, Deep learning-based solutions and, in particular, deep neural networks (DNNs) are getting ...
Deep learning-based solutions and, in particular, deep neural networks (DNNs) are at the heart of se...
Our work seeks to improve and adapt computing systems and machine learning (ML) algorithms to match ...
RISC-V is an open-source instruction set and now has been examined as a universal standard to unify ...
DNN inference is increasingly being executed locally on embedded platforms, due to the clear advanta...
Many applications such as autonomous driving and augmented reality, require the concurrent running o...
Deep neural network (DNN) inference is increasingly being executed on mobile and embedded platforms ...
Deep neural networks (DNNs) have emerged as successful solutions for variety of artificial intellige...
Thesis (Master's)--University of Washington, 2018Embedded platforms with integrated graphics process...
A plethora of applications are using machine learning, the operations of which are becoming more com...
GPUs are the workhorse in modern server infrastructure fueling advances in a number of compute-inten...
Deep Neural Network (DNN) inference is increasingly being deployed on edge devices, driven by the ad...
Part 2: AIInternational audienceConvolutional neural networks (CNNs) are widely used in vision-based...
Mobile and embedded platforms are increasingly required to efficiently execute computationally deman...
Platforms with multiple Field Programmable Gate Arrays (FPGAs), such as Amazon Web Services (AWS) F1...