This thesis discusses the application of optimizations to machine learning algorithms. In particular, we look at implementing these algorithms on specialized hardware, I.e. a Graphcore Intelligence Processing Unit, while also applying software optimizations that have been shown to improve performance of traditional workloads on general purpose CPUs. We discuss the feasibility of using these techniques when performing Matrix Factorization using Stochastic Gradient Descent on an IPU. We implement a program doing this, and show the results of changing different parameters during the running of SGD. We demonstrate that while machine learning is inherently approximate this does not mean that all approximate computation techniques are applicable,...
Modern learning problems in nature language processing, computer vision, computational biology, etc....
Machine learning has become one of the most exciting research areas in the world, with various appli...
This thesis proposes a novel experimental environment (non-linear stochastic gradient descent, NL-SG...
This thesis discusses the application of optimizations to machine learning algorithms. In particular...
Machine learning algorithms have opened up countless doors for scientists tackling problems that had...
This thesis aims at developing efficient optimization algorithms for solving large-scale machine lea...
Large scale machine learning requires tradeoffs. Commonly this tradeoff has led practitioners to cho...
Machine learning (ML) has been extensively employed for strategy optimization, decision making, data...
We propose an efficient machine learning algorithm for two-stage stochastic programs. This machine l...
Pervasive and networked computers have dramatically reduced the cost of collecting and distributing ...
On-line Machine Learning using Stochastic Gradient Descent is an inherently sequential computation. ...
Machine learning has been a topic in academia and industry for decades. Performance of machine lear...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
This thesis reports on experiments aimed at explaining why machine learning algorithms using the gre...
ML systems contend with an ever-growing processing load of physical world data. These systems are ...
Modern learning problems in nature language processing, computer vision, computational biology, etc....
Machine learning has become one of the most exciting research areas in the world, with various appli...
This thesis proposes a novel experimental environment (non-linear stochastic gradient descent, NL-SG...
This thesis discusses the application of optimizations to machine learning algorithms. In particular...
Machine learning algorithms have opened up countless doors for scientists tackling problems that had...
This thesis aims at developing efficient optimization algorithms for solving large-scale machine lea...
Large scale machine learning requires tradeoffs. Commonly this tradeoff has led practitioners to cho...
Machine learning (ML) has been extensively employed for strategy optimization, decision making, data...
We propose an efficient machine learning algorithm for two-stage stochastic programs. This machine l...
Pervasive and networked computers have dramatically reduced the cost of collecting and distributing ...
On-line Machine Learning using Stochastic Gradient Descent is an inherently sequential computation. ...
Machine learning has been a topic in academia and industry for decades. Performance of machine lear...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
This thesis reports on experiments aimed at explaining why machine learning algorithms using the gre...
ML systems contend with an ever-growing processing load of physical world data. These systems are ...
Modern learning problems in nature language processing, computer vision, computational biology, etc....
Machine learning has become one of the most exciting research areas in the world, with various appli...
This thesis proposes a novel experimental environment (non-linear stochastic gradient descent, NL-SG...