While deep neural networks (DNNs) have shown to be successful in several domains like computer vision, non-DNN models such as linear models and gradient boosting trees are still considered state-of-the-art over tabular data. When using these models, data scientists often author machine learning (ML) pipelines: DAG of ML operators comprising data transforms and ML models, whereby each operator is sequentially trained one-at-a-time. Conversely, when training DNNs, layers composing the neural networks are simultaneously trained using backpropagation.In this paper, we argue that the training scheme of ML pipelines is sub-optimal because it tries to optimize a single operator at a time thus losing the chance of global optimization. We therefore ...
Deep neural network architectures have recently produced excellent results in a variety of areas in ...
A multilayer perceptron is a feed forward artificial neural network model that maps sets of input da...
Thesis (Ph.D.)--University of Washington, 2016-06The choice of feature representation can have a lar...
While deep neural networks (DNNs) have shown to be successful in several domains like computer visio...
In this paper, we present our vision of differentiable ML pipelines called DiffML to automate the co...
Rapid progress in deep learning is leading to a diverse set of quickly changing models, with a drama...
Existing approaches that partition a convolutional neural network (CNN) onto multiple accelerators a...
Developments in deep learning with ANNs (Artificial Neural Networks) are paving the way for revoluti...
The growth in size and complexity of convolutional neural networks (CNNs) is forcing the partitionin...
Multilayer Neural Networks (MNNs) are commonly trained using gradient descent-based methods, such as...
In recent years, machine learning (ML) and, more noticeably, deep learning (DL), have be- come incre...
Supervised Learning in Multi-Layered Neural Networks (MLNs) has been recently proposed through the w...
It is possible to learn multiple layers of non-linear features by backpropa-gating error derivatives...
We present DANTE, a novel method for training neural networks using the alternating minimization pri...
Artificial Neural Networks (ANNs) are often used (trained) to find a general solution in problems wh...
Deep neural network architectures have recently produced excellent results in a variety of areas in ...
A multilayer perceptron is a feed forward artificial neural network model that maps sets of input da...
Thesis (Ph.D.)--University of Washington, 2016-06The choice of feature representation can have a lar...
While deep neural networks (DNNs) have shown to be successful in several domains like computer visio...
In this paper, we present our vision of differentiable ML pipelines called DiffML to automate the co...
Rapid progress in deep learning is leading to a diverse set of quickly changing models, with a drama...
Existing approaches that partition a convolutional neural network (CNN) onto multiple accelerators a...
Developments in deep learning with ANNs (Artificial Neural Networks) are paving the way for revoluti...
The growth in size and complexity of convolutional neural networks (CNNs) is forcing the partitionin...
Multilayer Neural Networks (MNNs) are commonly trained using gradient descent-based methods, such as...
In recent years, machine learning (ML) and, more noticeably, deep learning (DL), have be- come incre...
Supervised Learning in Multi-Layered Neural Networks (MLNs) has been recently proposed through the w...
It is possible to learn multiple layers of non-linear features by backpropa-gating error derivatives...
We present DANTE, a novel method for training neural networks using the alternating minimization pri...
Artificial Neural Networks (ANNs) are often used (trained) to find a general solution in problems wh...
Deep neural network architectures have recently produced excellent results in a variety of areas in ...
A multilayer perceptron is a feed forward artificial neural network model that maps sets of input da...
Thesis (Ph.D.)--University of Washington, 2016-06The choice of feature representation can have a lar...