Published April 2019 | Published + Submitted
Journal Article Open

An Online Algorithm for Smoothed Regression and LQR Control

An error occurred while generating the citation.

Abstract

We consider Online Convex Optimization (OCO) in the setting where the costs are mm-strongly convex and the online learner pays a switching cost for changing decisions between rounds. We show that the recently proposed Online Balanced Descent (OBD) algorithm is constant competitive in this setting, with competitive ratio 3+O(1/m), irrespective of the ambient dimension. Additionally, we show that when the sequence of cost functions is ϵϵ-smooth, OBD has near-optimal dynamic regret and maintains strong per-round accuracy. We demonstrate the generality of our approach by showing that the OBD framework can be used to construct competitive algorithms for a variety of online problems across learning and control, including online variants of ridge regression, logistic regression, maximum likelihood estimation, and LQR control.

Additional Information

© 2019 by the author(s).

Attached Files

Published - goel19a.pdf

Submitted - 1810.10132.pdf

Files

goel19a.pdf
Files (637.5 kB)
Name Size Download all
md5:46c3d7699fc1d465e516d359835c8808
292.3 kB Preview Download
md5:4cb79059f8aafedd07769f3986882652
345.3 kB Preview Download

Additional details

Created:
August 19, 2023
Modified:
October 20, 2023