Online strongly convex optimization with unknown delays

作者:Yuanyu Wan, Wei-Wei Tu, Lijun Zhang

摘要

We investigate the problem of online convex optimization with unknown delays, in which the feedback of a decision arrives with an arbitrary delay. Previous studies have presented delayed online gradient descent (DOGD), and achieved the regret bound of \(O(\sqrt{D})\) by only utilizing the convexity condition, where \(D\ge T\) is the sum of delays over T rounds. In this paper, we further exploit the strong convexity to improve the regret bound. Specifically, we first propose a variant of DOGD for strongly convex functions, and establish a better regret bound of \(O(d\log T)\), where d is the maximum delay. The essential idea is to let the learning rate decay with the total number of received feedback linearly. Furthermore, we extend the strongly convex variant of DOGD and its theoretical guarantee to the more challenging bandit setting by combining with the classical \((n+1)\)-point and two-point gradient estimators, where n is the dimensionality. To the best of our knowledge, this is the first work that solves online strongly convex optimization under the general delayed setting.

论文关键词:Online optimization, Strongly convex, Unknown delays, Bandit

论文评审过程:

论文官网地址:https://doi.org/10.1007/s10994-021-06072-w