Welcome to the new version of CaltechAUTHORS. Login is currently restricted to library staff. If you notice any issues, please email coda@library.caltech.edu
Published April 2019 | Published + Submitted
Journal Article Open

An Online Algorithm for Smoothed Regression and LQR Control

Abstract

We consider Online Convex Optimization (OCO) in the setting where the costs are mm-strongly convex and the online learner pays a switching cost for changing decisions between rounds. We show that the recently proposed Online Balanced Descent (OBD) algorithm is constant competitive in this setting, with competitive ratio 3+O(1/m), irrespective of the ambient dimension. Additionally, we show that when the sequence of cost functions is ϵϵ-smooth, OBD has near-optimal dynamic regret and maintains strong per-round accuracy. We demonstrate the generality of our approach by showing that the OBD framework can be used to construct competitive algorithms for a variety of online problems across learning and control, including online variants of ridge regression, logistic regression, maximum likelihood estimation, and LQR control.

Additional Information

© 2019 by the author(s).

Attached Files

Published - goel19a.pdf

Submitted - 1810.10132.pdf

Files

goel19a.pdf
Files (637.5 kB)
Name Size Download all
md5:46c3d7699fc1d465e516d359835c8808
292.3 kB Preview Download
md5:4cb79059f8aafedd07769f3986882652
345.3 kB Preview Download

Additional details

Created:
August 19, 2023
Modified:
October 20, 2023