An Online Algorithm for Smoothed Regression and LQR Control
- Creators
-
Goel, Gautam
-
Wierman, Adam
Abstract
We consider Online Convex Optimization (OCO) in the setting where the costs are mm-strongly convex and the online learner pays a switching cost for changing decisions between rounds. We show that the recently proposed Online Balanced Descent (OBD) algorithm is constant competitive in this setting, with competitive ratio 3+O(1/m), irrespective of the ambient dimension. Additionally, we show that when the sequence of cost functions is ϵϵ-smooth, OBD has near-optimal dynamic regret and maintains strong per-round accuracy. We demonstrate the generality of our approach by showing that the OBD framework can be used to construct competitive algorithms for a variety of online problems across learning and control, including online variants of ridge regression, logistic regression, maximum likelihood estimation, and LQR control.
Additional Information
© 2019 by the author(s).Attached Files
Published - goel19a.pdf
Submitted - 1810.10132.pdf
Files
Name | Size | Download all |
---|---|---|
md5:46c3d7699fc1d465e516d359835c8808
|
292.3 kB | Preview Download |
md5:4cb79059f8aafedd07769f3986882652
|
345.3 kB | Preview Download |
Additional details
- Alternative title
- Smoothed Online Optimization for Regression and Control
- Eprint ID
- 96758
- Resolver ID
- CaltechAUTHORS:20190626-160602759
- Created
-
2019-06-27Created from EPrint's datestamp field
- Updated
-
2023-06-02Created from EPrint's last_modified field