Stochastic and adversarial data are two widely studied settings in online learning. But many optimization tasks are neither i.i.d. nor fully adversarial, which makes it of fundamental interest to get a better theoretical understanding of the world between these extremes. In this work we establish novel regret bounds for online convex optimization in a setting that interpolates between stochastic i.i.d. and fully adversarial losses. By exploiting smoothness of the expected losses, these bounds replace a dependence on the maximum gradient length by the variance of the gradients, which was previously known only for linear losses. In addition, they weaken the i.i.d. assumption by allowing, for example, adversarially poisoned rounds, which were ...
We demonstrate a modification of the algorithm of Dani et al for the online linear optimization prob...
The framework of online learning with memory naturally captures learning problems with temporal effe...
Much of the work in online learning focuses on the study of sublinear upper bounds on the regret. In...
The study of online convex optimization in the bandit setting was initiated by Klein-berg (2004) and...
We study the regret of optimal strategies for online convex optimization games. Using von Neumann's ...
We consider online convex optimizations in the bandit setting. The decision maker does not know the ...
We study the regret of optimal strategies for online convex optimization games. Using von Neumann's ...
We consider the online convex optimization problem. In the setting of arbitrary sequences and finite...
Bandit convex optimization is a special case of online convex optimization with partial information....
We consider the online convex optimization problem. In the setting of arbitrary sequences and finite...
The regret bound of dynamic online learning algorithms is often expressed in terms of the variation ...
Consider the online convex optimization problem, in which a player has to choose ac-tions iterativel...
We study the rates of growth of the regret in online convex optimization. First, we show that a simp...
We aim to design universal algorithms for online convex optimization, which can handle multiple comm...
We aim to design universal algorithms for online convex optimization, which can handle multiple comm...
We demonstrate a modification of the algorithm of Dani et al for the online linear optimization prob...
The framework of online learning with memory naturally captures learning problems with temporal effe...
Much of the work in online learning focuses on the study of sublinear upper bounds on the regret. In...
The study of online convex optimization in the bandit setting was initiated by Klein-berg (2004) and...
We study the regret of optimal strategies for online convex optimization games. Using von Neumann's ...
We consider online convex optimizations in the bandit setting. The decision maker does not know the ...
We study the regret of optimal strategies for online convex optimization games. Using von Neumann's ...
We consider the online convex optimization problem. In the setting of arbitrary sequences and finite...
Bandit convex optimization is a special case of online convex optimization with partial information....
We consider the online convex optimization problem. In the setting of arbitrary sequences and finite...
The regret bound of dynamic online learning algorithms is often expressed in terms of the variation ...
Consider the online convex optimization problem, in which a player has to choose ac-tions iterativel...
We study the rates of growth of the regret in online convex optimization. First, we show that a simp...
We aim to design universal algorithms for online convex optimization, which can handle multiple comm...
We aim to design universal algorithms for online convex optimization, which can handle multiple comm...
We demonstrate a modification of the algorithm of Dani et al for the online linear optimization prob...
The framework of online learning with memory naturally captures learning problems with temporal effe...
Much of the work in online learning focuses on the study of sublinear upper bounds on the regret. In...