Black-box optimization is primarily important for many computationally intensive applications, including reinforcement learning (RL), robot control, etc. This paper presents a novel theoretical framework for black-box optimization, in which our method performs stochastic updates with an implicit natural gradient of an exponential-family distribution. Theoretically, we prove the convergence rate of our framework with full matrix update for convex functions under Gaussian distribution. Our methods are very simple and contain fewer hyper-parameters than CMA-ES [12]. Empirically, our method with full matrix update achieves competitive performance compared with one of the state-of-the-art methods CMA-ES on benchmark test problems. Moreover, our ...
We consider stochastic second-order methods for minimizing smooth and strongly-convex functions unde...
The diverse world of machine learning applications has given rise to a plethora of algorithms and op...
International audienceIn this paper we investigate the convergence properties of a variant of the Co...
International audienceThis work considers stochastic optimization problems in which the objective fu...
Optimization of black-box functions has been of interest to researchers for many years and has beco...
This monograph presents the main mathematical ideas in convex opti-mization. Starting from the funda...
We develop an optimization algorithm suitable for Bayesian learning in complex models. Our approach ...
The goal of this paper is to debunk and dispel the magic behind black-box optimizers and stochastic ...
31 pages, 4 figures, 1 tableInternational audienceUniversal methods for optimization are designed to...
University of Technology Sydney. Faculty of Engineering and Information Technology.Black-box optimiz...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
Gradient-free/zeroth-order methods for black-box convex optimization have been extensively studied i...
Regularized risk minimization often involves non-smooth optimization, either because of the loss fun...
National audienceThis paper studies some asymptotic properties of adaptive algorithms widely used in...
Abstract. Linear optimization is many times algorithmically simpler than non-linear convex optimizat...
We consider stochastic second-order methods for minimizing smooth and strongly-convex functions unde...
The diverse world of machine learning applications has given rise to a plethora of algorithms and op...
International audienceIn this paper we investigate the convergence properties of a variant of the Co...
International audienceThis work considers stochastic optimization problems in which the objective fu...
Optimization of black-box functions has been of interest to researchers for many years and has beco...
This monograph presents the main mathematical ideas in convex opti-mization. Starting from the funda...
We develop an optimization algorithm suitable for Bayesian learning in complex models. Our approach ...
The goal of this paper is to debunk and dispel the magic behind black-box optimizers and stochastic ...
31 pages, 4 figures, 1 tableInternational audienceUniversal methods for optimization are designed to...
University of Technology Sydney. Faculty of Engineering and Information Technology.Black-box optimiz...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
Gradient-free/zeroth-order methods for black-box convex optimization have been extensively studied i...
Regularized risk minimization often involves non-smooth optimization, either because of the loss fun...
National audienceThis paper studies some asymptotic properties of adaptive algorithms widely used in...
Abstract. Linear optimization is many times algorithmically simpler than non-linear convex optimizat...
We consider stochastic second-order methods for minimizing smooth and strongly-convex functions unde...
The diverse world of machine learning applications has given rise to a plethora of algorithms and op...
International audienceIn this paper we investigate the convergence properties of a variant of the Co...