Adaptive Online Non-stochastic Control

Research output: Chapter in Book/Conference proceedings/Edited volumeConference contributionScientificpeer-review

27 Downloads (Pure)

Abstract

We tackle the problem of Non-stochastic Control (NSC) with the aim of obtaining algorithms whose policy regret is proportional to the difficulty of the controlled environment. Namely, we tailor the Follow The Regularized Leader (FTRL) framework to dynamical systems by using regularizers that are proportional to the actual witnessed costs. The main challenge arises from using the proposed adaptive regularizers in the presence of a state, or equivalently, a memory, which couples the effect of the online decisions and requires new tools for bounding the regret. Via new analysis techniques for NSC and FTRL integration, we obtain novel disturbance action controllers (DAC) with sub-linear data adaptive policy regret bounds that shrink when the trajectory of costs has small gradients, while staying sub-linear even in the worst case.

Original languageEnglish
Title of host publicationProceedings of Machine Learning Research
EditorsA. Abate, M. Cannon, K. Margellos, A. Papachristodoulou
Pages248-259
Number of pages12
Volume242
Publication statusPublished - 2024
Event6th Annual Learning for Dynamics and Control Conference, L4DC 2024 - Oxford, United Kingdom
Duration: 15 Jul 202417 Jul 2024

Publication series

NameProceedings of Machine Learning Research
ISSN (Print)2640-3498

Conference

Conference6th Annual Learning for Dynamics and Control Conference, L4DC 2024
Country/TerritoryUnited Kingdom
CityOxford
Period15/07/2417/07/24

Keywords

  • Follow the Regularized Leader
  • Non-stochastic control
  • Online learning

Fingerprint

Dive into the research topics of 'Adaptive Online Non-stochastic Control'. Together they form a unique fingerprint.

Cite this