In machine learning research, many emerging applications can be (re)formulated as the composition optimization problem with nonsmooth regularization penalty. To solve this problem, traditional stochastic gradient descent (SGD) algorithm and its variants either have low convergence rate or are computationally expensive. Recently, several stochastic composition gradient algorithms have been proposed, however, these methods are still inefficient and not scalable to large-scale composition optimization problem instances. To address these challenges, we propose an asynchronous parallel algorithm, named Async-ProxSCVR, which effectively combines asynchronous parallel implementation and variance reduction method. We prove that the algorithm admits...
By combining randomized smoothing techniques with accelerated gradient methods, we obtain convergenc...
Recent studies have illustrated that stochastic gradient Markov Chain Monte Carlo techniques have a ...
By combining randomized smoothing techniques with accelerated gradient methods, we obtain convergenc...
Consider the stochastic composition optimization problem where the objective is a composition of two...
Stochastic composition optimization draws much attention recently and has been successful in many e...
We provide the first theoretical analysis on the convergence rate of asynchronous mini-batch gradie...
Stochastic gradient descent (SGD) and its variants have become more and more popular in machine lear...
Regularized empirical risk minimization (R-ERM) is an important branch of machine learning, since it...
The implementation of a vast majority of machine learning (ML) algorithms boils down to solving a nu...
Nowadays, asynchronous parallel algorithms have received much attention in the optimization field du...
The implementation of a vast majority of machine learning (ML) algorithms boils down to solving a nu...
With the recent proliferation of large-scale learning problems, there have been a lot of interest o...
Stochastic Gradient Descent (SGD) is very useful in optimization problems with high-dimensional non-...
Recent studies have illustrated that stochastic gradient Markov Chain Monte Carlo techniques have a ...
By combining randomized smoothing techniques with accelerated gradient methods, we obtain convergenc...
By combining randomized smoothing techniques with accelerated gradient methods, we obtain convergenc...
Recent studies have illustrated that stochastic gradient Markov Chain Monte Carlo techniques have a ...
By combining randomized smoothing techniques with accelerated gradient methods, we obtain convergenc...
Consider the stochastic composition optimization problem where the objective is a composition of two...
Stochastic composition optimization draws much attention recently and has been successful in many e...
We provide the first theoretical analysis on the convergence rate of asynchronous mini-batch gradie...
Stochastic gradient descent (SGD) and its variants have become more and more popular in machine lear...
Regularized empirical risk minimization (R-ERM) is an important branch of machine learning, since it...
The implementation of a vast majority of machine learning (ML) algorithms boils down to solving a nu...
Nowadays, asynchronous parallel algorithms have received much attention in the optimization field du...
The implementation of a vast majority of machine learning (ML) algorithms boils down to solving a nu...
With the recent proliferation of large-scale learning problems, there have been a lot of interest o...
Stochastic Gradient Descent (SGD) is very useful in optimization problems with high-dimensional non-...
Recent studies have illustrated that stochastic gradient Markov Chain Monte Carlo techniques have a ...
By combining randomized smoothing techniques with accelerated gradient methods, we obtain convergenc...
By combining randomized smoothing techniques with accelerated gradient methods, we obtain convergenc...
Recent studies have illustrated that stochastic gradient Markov Chain Monte Carlo techniques have a ...
By combining randomized smoothing techniques with accelerated gradient methods, we obtain convergenc...