no code implementations • 12 Dec 2023 • Gautam Goel, Peter Bartlett
We revisit the problem of Kalman Filtering in linear dynamical systems and show that Transformers can approximate the Kalman Filter in a strong sense.
no code implementations • 21 Nov 2022 • Gautam Goel, Naman Agarwal, Karan Singh, Elad Hazan
We consider the fundamental problem of online control of a linear dynamical system from two different viewpoints: regret minimization and competitive analysis.
no code implementations • 24 Oct 2021 • Gautam Goel, Babak Hassibi
A natural goal when designing online learning algorithms for non-stationary environments is to bound the regret of the algorithm in terms of the temporal variation of the input sequence.
no code implementations • 28 Jul 2021 • Gautam Goel, Babak Hassibi
We consider control from the perspective of competitive analysis.
no code implementations • 22 Jun 2021 • Gautam Goel, Babak Hassibi
We consider estimation and control in linear time-varying dynamical systems from the perspective of regret minimization.
no code implementations • 4 May 2021 • Oron Sabag, Gautam Goel, Sahin Lale, Babak Hassibi
Motivated by competitive analysis in online learning, as a criterion for controller design we introduce the dynamic regret, defined as the difference between the LQR cost of a causal controller (that has only access to past disturbances) and the LQR cost of the \emph{unique} clairvoyant one (that has also access to future disturbances) that is known to dominate all other controllers.
no code implementations • 24 Nov 2020 • Gautam Goel, Babak Hassibi
We consider measurement-feedback control in linear dynamical systems from the perspective of regret minimization.
no code implementations • 20 Oct 2020 • Gautam Goel, Babak Hassibi
We consider control in linear time-varying dynamical systems from the perspective of regret minimization.
no code implementations • 7 Feb 2020 • Gautam Goel, Babak Hassibi
We also show that cost of the optimal offline linear policy converges to the cost of the optimal online policy as the time horizon grows large, and consequently the optimal offline linear policy incurs linear regret relative to the optimal offline policy, even in the optimistic setting where the noise is drawn i. i. d from a known distribution.
no code implementations • 10 Nov 2019 • Yiheng Lin, Gautam Goel, Adam Wierman
In this work, we give two general sufficient conditions that specify a relationship between the hitting and movement costs which guarantees that a new algorithm, Synchronized Fixed Horizon Control (SFHC), provides a $1+O(1/w)$ competitive ratio, where $w$ is the number of predictions available to the learner.
no code implementations • NeurIPS 2019 • Gautam Goel, Yiheng Lin, Haoyuan Sun, Adam Wierman
We prove a new lower bound on the competitive ratio of any online algorithm in the setting where the costs are $m$-strongly convex and the movement costs are the squared $\ell_2$ norm.
no code implementations • 23 Oct 2018 • Gautam Goel, Adam Wierman
We consider Online Convex Optimization (OCO) in the setting where the costs are $m$-strongly convex and the online learner pays a switching cost for changing decisions between rounds.
no code implementations • 28 Mar 2018 • Niangjun Chen, Gautam Goel, Adam Wierman
We demonstrate the generality of the OBD framework by showing how, with different choices of "balance," OBD can improve upon state-of-the-art performance guarantees for both competitive ratio and regret, in particular, OBD is the first algorithm to achieve a dimension-free competitive ratio, $3 + O(1/\alpha)$, for locally polyhedral costs, where $\alpha$ measures the "steepness" of the costs.