We consider stochastic second-order methods for minimizing smooth and strongly-convex functions under an interpolation condition, which can be satisfied by over-parameterized machine learning models. Under this condition, we show that the regularized subsampled Newton’s method (R-SSN) achieves global linear convergence with an adaptive step-size and a constant batch-size. By growing the batch size for both the subsampled gradient and Hessian, we show that R-SSN can converge at a quadratic rate in a local neighbourhood of the solution. We also show that R-SSN attains local linear convergence for the family of self-concordant functions. Furthermore, we analyze stochastic BFGS algorithms in the interpolation setting and prove their global line...
In this dissertation, we propose two stochastic alternating optimization methods for solving struct...
This paper proposes a new family of algorithms for training neural networks (NNs). These...
Large-scale machine learning problems can be reduced to non-convex optimization problems if state-of...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
We study the performance of Stochastic Cubic Regularized Newton (SCRN) on a class of functions satis...
Stochastic optimization has received extensive attention in recent years due to their extremely pote...
We develop stochastic variants of the wellknown BFGS quasi-Newton optimization method, in both full ...
We study the convergence, the implicit regularization and the generalization of stochastic mirror de...
We study the convergence, the implicit regularization and the generalization of stochastic mirror de...
While first-order methods are popular for solving optimization problems that arise in large-scale de...
Recent years have witnessed huge advances in machine learning (ML) and its applications, especially ...
Recent years have witnessed huge advances in machine learning (ML) and its applications, especially ...
Large-scale machine learning problems can be reduced to non-convex optimization problems if state-of...
This work considers optimization methods for large-scale machine learning (ML). Optimization in ML ...
In this dissertation, we propose two stochastic alternating optimization methods for solving struct...
This paper proposes a new family of algorithms for training neural networks (NNs). These...
Large-scale machine learning problems can be reduced to non-convex optimization problems if state-of...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
We study the performance of Stochastic Cubic Regularized Newton (SCRN) on a class of functions satis...
Stochastic optimization has received extensive attention in recent years due to their extremely pote...
We develop stochastic variants of the wellknown BFGS quasi-Newton optimization method, in both full ...
We study the convergence, the implicit regularization and the generalization of stochastic mirror de...
We study the convergence, the implicit regularization and the generalization of stochastic mirror de...
While first-order methods are popular for solving optimization problems that arise in large-scale de...
Recent years have witnessed huge advances in machine learning (ML) and its applications, especially ...
Recent years have witnessed huge advances in machine learning (ML) and its applications, especially ...
Large-scale machine learning problems can be reduced to non-convex optimization problems if state-of...
This work considers optimization methods for large-scale machine learning (ML). Optimization in ML ...
In this dissertation, we propose two stochastic alternating optimization methods for solving struct...
This paper proposes a new family of algorithms for training neural networks (NNs). These...
Large-scale machine learning problems can be reduced to non-convex optimization problems if state-of...