Big Data problems in Machine Learning have large number of data points or large number of features, or both, which make training of models difficult because of high computational complexities of single iteration of learning algorithms. To solve such learning problems, Stochastic Approximation offers an optimization approach to make complexity of each iteration independent of number of data points by taking only one data point or mini-batch of data points during each iteration and thereby helping to solve problems with large number of data points. Similarly, Coordinate Descent offers another optimization approach to make iteration complexity independent of the number of features/coordinates/variables by taking only one feature or block of fe...
In this work we show that randomized (block) coordinate descent methods can be accelerated by parall...
The steplength selection is a crucial issue for the effectiveness of the stochastic gradient methods...
Two types of low cost-per-iteration gradient descent methods have been extensively studied in par-al...
big data optimization in machine learning: special structure Single machine optimization stochastic ...
Stochastic gradient descent (SGD) holds as a classical method to build large scale machine learning ...
The stochastic gradient (SG) method can minimize an objective function composed of a large number of...
Abstract. The stochastic gradient (SG) method can quickly solve a problem with a large number of com...
The stochastic gradient (SG) method can minimize an objective function composed of a large number of...
Recent years have witnessed huge advances in machine learning (ML) and its applications, especially ...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
University of Minnesota Ph.D. dissertation. April 2020. Major: Computer Science. Advisor: Arindam Ba...
Thesis (Ph.D.)--University of Washington, 2019Tremendous advances in large scale machine learning an...
Stochastic optimization has received extensive attention in recent years due to their extremely pote...
The unprecedented rate at which data is being created and stored calls for scalable optimization te...
Gradient-based algorithms are popular when solving unconstrained optimization problems. By exploitin...
In this work we show that randomized (block) coordinate descent methods can be accelerated by parall...
The steplength selection is a crucial issue for the effectiveness of the stochastic gradient methods...
Two types of low cost-per-iteration gradient descent methods have been extensively studied in par-al...
big data optimization in machine learning: special structure Single machine optimization stochastic ...
Stochastic gradient descent (SGD) holds as a classical method to build large scale machine learning ...
The stochastic gradient (SG) method can minimize an objective function composed of a large number of...
Abstract. The stochastic gradient (SG) method can quickly solve a problem with a large number of com...
The stochastic gradient (SG) method can minimize an objective function composed of a large number of...
Recent years have witnessed huge advances in machine learning (ML) and its applications, especially ...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
University of Minnesota Ph.D. dissertation. April 2020. Major: Computer Science. Advisor: Arindam Ba...
Thesis (Ph.D.)--University of Washington, 2019Tremendous advances in large scale machine learning an...
Stochastic optimization has received extensive attention in recent years due to their extremely pote...
The unprecedented rate at which data is being created and stored calls for scalable optimization te...
Gradient-based algorithms are popular when solving unconstrained optimization problems. By exploitin...
In this work we show that randomized (block) coordinate descent methods can be accelerated by parall...
The steplength selection is a crucial issue for the effectiveness of the stochastic gradient methods...
Two types of low cost-per-iteration gradient descent methods have been extensively studied in par-al...