Search Results for author: Jonathan Wilder Lavington

Found 10 papers, 4 papers with code

Analyzing and Improving Greedy 2-Coordinate Updates for Equality-Constrained Optimization via Steepest Descent in the 1-Norm

no code implementations3 Jul 2023 Amrutha Varshini Ramesh, Aaron Mishkin, Mark Schmidt, Yihan Zhou, Jonathan Wilder Lavington, Jennifer She

We show that bound- and summation-constrained steepest descent in the L1-norm guarantees more progress per iteration than previous rules and can be computed in only $O(n \log n)$ time.

Video Killed the HD-Map: Predicting Multi-Agent Behavior Directly From Aerial Images

no code implementations19 May 2023 Yunpeng Liu, Vasileios Lioutas, Jonathan Wilder Lavington, Matthew Niedoba, Justice Sefas, Setareh Dabiri, Dylan Green, Xiaoxuan Liang, Berend Zwartsenberg, Adam Ścibior, Frank Wood

The development of algorithms that learn multi-agent behavioral models using human demonstrations has led to increasingly realistic simulations in the field of autonomous driving.

Autonomous Driving Trajectory Prediction

Target-based Surrogates for Stochastic Optimization

1 code implementation6 Feb 2023 Jonathan Wilder Lavington, Sharan Vaswani, Reza Babanezhad, Mark Schmidt, Nicolas Le Roux

Our target optimization framework uses the (expensive) gradient computation to construct surrogate functions in a \emph{target space} (e. g. the logits output by a linear model for classification) that can be minimized efficiently.

Imitation Learning Stochastic Optimization

Vehicle Type Specific Waypoint Generation

no code implementations9 Aug 2022 Yunpeng Liu, Jonathan Wilder Lavington, Adam Scibior, Frank Wood

We develop a generic mechanism for generating vehicle-type specific sequences of waypoints from a probabilistic foundation model of driving behavior.

reinforcement-learning Reinforcement Learning (RL) +1

Improved Policy Optimization for Online Imitation Learning

1 code implementation29 Jul 2022 Jonathan Wilder Lavington, Sharan Vaswani, Mark Schmidt

Specifically, if the class of policies is sufficiently expressive to contain the expert policy, we prove that DAGGER achieves constant regret.

Imitation Learning

Critic Sequential Monte Carlo

no code implementations30 May 2022 Vasileios Lioutas, Jonathan Wilder Lavington, Justice Sefas, Matthew Niedoba, Yunpeng Liu, Berend Zwartsenberg, Setareh Dabiri, Frank Wood, Adam Scibior

We introduce CriticSMC, a new algorithm for planning as inference built from a composition of sequential Monte Carlo with learned Soft-Q function heuristic factors.

Collision Avoidance

Cannot find the paper you are looking for? You can Submit a new open access paper.