International audienceWe present the novel idea of inference delivery networks (IDN), networks of computing nodes that coordinate to satisfy inference requests achieving the best trade-off between latency and accuracy. IDNs bridge the dichotomy between device and cloud execution by integrating inference delivery at the various tiers of the infrastructure continuum (access, edge, regional data center, cloud). We propose a distributed dynamic policy for ML model allocation in an IDN by which each node periodically updates its local set of inference models based on requests observed during the recent past plus limited information exchange with its neighbor nodes. Our policy offers strong performance guarantees in an adversarial setting and sho...
Natural Language Inference (NLI) is a key, complex task where machine learning (ML) is playing an im...
The increasingly growing expansion of the Internet of Things (IoT) along with the convergence of mul...
We address distributed machine learning in multi-tier (e.g., mobile-edge-cloud) networks where a het...
With the advancement of machine learning, a growing number of mobile users rely on machine learning ...
Inference carried out on pre-trained deep neural networks (DNNs) is particularly effective as it doe...
With the advent of ubiquitous deployment of smart devices and the Internet of Things, data sources f...
Thesis (Master's)--University of Washington, 2021With the advancement of machine learning (ML), a gr...
Deep neural networks (DNN) are the de-facto solution behind many intelligent applications of today, ...
While high accuracy is of paramount importance for deep learning (DL) inference, serving inference r...
National audienceThe emergence of Machine Learning (ML) has increased exponentially in numerous appl...
With the rise of machine learning, inference on deep neural networks (DNNs) has become a core buildi...
In today’s world machine learning has major applications in a wide variety of tasks such as image c...
Deep Neural Networks (DNNs) based on intelligent applications have been intensively deployed on mobi...
The success of deep neural networks (DNNs) is heavily dependent on computational resources. While DN...
We study distributed inference, learning and optimization in scenarios which involve networked entit...
Natural Language Inference (NLI) is a key, complex task where machine learning (ML) is playing an im...
The increasingly growing expansion of the Internet of Things (IoT) along with the convergence of mul...
We address distributed machine learning in multi-tier (e.g., mobile-edge-cloud) networks where a het...
With the advancement of machine learning, a growing number of mobile users rely on machine learning ...
Inference carried out on pre-trained deep neural networks (DNNs) is particularly effective as it doe...
With the advent of ubiquitous deployment of smart devices and the Internet of Things, data sources f...
Thesis (Master's)--University of Washington, 2021With the advancement of machine learning (ML), a gr...
Deep neural networks (DNN) are the de-facto solution behind many intelligent applications of today, ...
While high accuracy is of paramount importance for deep learning (DL) inference, serving inference r...
National audienceThe emergence of Machine Learning (ML) has increased exponentially in numerous appl...
With the rise of machine learning, inference on deep neural networks (DNNs) has become a core buildi...
In today’s world machine learning has major applications in a wide variety of tasks such as image c...
Deep Neural Networks (DNNs) based on intelligent applications have been intensively deployed on mobi...
The success of deep neural networks (DNNs) is heavily dependent on computational resources. While DN...
We study distributed inference, learning and optimization in scenarios which involve networked entit...
Natural Language Inference (NLI) is a key, complex task where machine learning (ML) is playing an im...
The increasingly growing expansion of the Internet of Things (IoT) along with the convergence of mul...
We address distributed machine learning in multi-tier (e.g., mobile-edge-cloud) networks where a het...