13/04/2021

Power of hints for online learning with movement costs

Aditya Bhaskara, Ashok Cutkosky, Ravi Kumar, Manish Purohit

Keywords:

Abstract: We consider the online linear optimization problem with movement costs, a variant of online learning in which the learner must not only respond to cost vectors c_t with points x_t in order to maintain low regret, but is also penalized for movement by an additional cost \|x_t-x_{t+1}\|^{1+\epsilon} for some \epsilon>0. Classically, simple algorithms that obtain the optimal \sqrt{T} regret already are very stable and do not incur a significant movement cost. However, recent work has shown that when the learning algorithm is provided with weak “hint” vectors that have a positive correlation with the costs, the regret can be significantly improved to \log(T). In this work, we study the stability of such algorithms, and provide matching upper and lower bounds showing that incorporating movement costs results in intricate tradeoffs between \log(T) when \epsilon\ge 1 and \sqrt{T} regret when \epsilon=0.

 0
 0
 0
 0
This is an embedded video. Talk and the respective paper are published at AISTATS 2021 virtual conference. If you are one of the authors of the paper and want to manage your upload, see the question "My papertalk has been externally embedded..." in the FAQ section.

Comments

Post Comment
no comments yet
code of conduct: tbd Characters remaining: 140

Similar Papers