abstract: This thesis presents a family of adaptive curvature methods for gradient-based stochastic optimization. In particular, a general algorithmic framework is introduced along with a practical implementation that yields an efficient, adaptive curvature gradient descent algorithm. To this end, a theoretical and practical link between curvature matrix estimation and shrinkage methods for covariance matrices is established. The use of shrinkage improves estimation accuracy of the curvature matrix when data samples are scarce. This thesis also introduce several insights that result in data- and computation-efficient update equations. Empirical results suggest that the proposed method compares favorably with existing second-order techniques...
We introduce adaptive sampling methods for stochastic programs with deterministic constraints. First...
International audienceIn this paper we investigate the convergence properties of a variant of the Co...
237 pagesIt seems that in the current age, computers, computation, and data have an increasingly imp...
The emergent field of machine learning has by now become the main proponent of data-driven discovery...
Incorporating second-order curvature information into machine learning optimization algorithms can b...
This paper introduces PROMISE ($\textbf{Pr}$econditioned Stochastic $\textbf{O}$ptimization $\textbf...
where the function f or its gradient rf are not directly accessible except through Monte Carlo estim...
In this thesis, we develop a novel and efficient algorithm for optimizing neural networks inspired b...
In this paper we propose several adaptive gradient methods for stochastic optimization. Our methods ...
Treballs finals del Màster en Matemàtica Avançada, Facultat de matemàtiques, Universitat de Barcelon...
In this dissertation, a theoretical framework based on concentration inequalities for empirical proc...
Gradient-based optimization and Markov Chain Monte Carlo sampling can be found at the heart of sever...
In this work, we introduce AdaCN, a novel adaptive cubic Newton method for nonconvex stochastic opti...
This thesis aims at developing efficient optimization algorithms for solving large-scale machine lea...
Gradient-based optimization algorithms, in particular their stochastic counterparts, have become by ...
We introduce adaptive sampling methods for stochastic programs with deterministic constraints. First...
International audienceIn this paper we investigate the convergence properties of a variant of the Co...
237 pagesIt seems that in the current age, computers, computation, and data have an increasingly imp...
The emergent field of machine learning has by now become the main proponent of data-driven discovery...
Incorporating second-order curvature information into machine learning optimization algorithms can b...
This paper introduces PROMISE ($\textbf{Pr}$econditioned Stochastic $\textbf{O}$ptimization $\textbf...
where the function f or its gradient rf are not directly accessible except through Monte Carlo estim...
In this thesis, we develop a novel and efficient algorithm for optimizing neural networks inspired b...
In this paper we propose several adaptive gradient methods for stochastic optimization. Our methods ...
Treballs finals del Màster en Matemàtica Avançada, Facultat de matemàtiques, Universitat de Barcelon...
In this dissertation, a theoretical framework based on concentration inequalities for empirical proc...
Gradient-based optimization and Markov Chain Monte Carlo sampling can be found at the heart of sever...
In this work, we introduce AdaCN, a novel adaptive cubic Newton method for nonconvex stochastic opti...
This thesis aims at developing efficient optimization algorithms for solving large-scale machine lea...
Gradient-based optimization algorithms, in particular their stochastic counterparts, have become by ...
We introduce adaptive sampling methods for stochastic programs with deterministic constraints. First...
International audienceIn this paper we investigate the convergence properties of a variant of the Co...
237 pagesIt seems that in the current age, computers, computation, and data have an increasingly imp...