Search Results for author: Joel W. Burdick

Found 33 papers, 7 papers with code

Disturbance Observers for Robust Backup Control Barrier Functions

no code implementations19 Mar 2025 David E. J. van Wijk, Ersin Das, Anil Alan, Samuel Coogan, Tamas G. Molnar, Joel W. Burdick, Manoranjan Majji, Kerianne L. Hobbs

To tackle this shortcoming, we integrate backup control barrier functions with a disturbance observer and estimate the flow under a reconstruction of the disturbance while refining this estimate over time.

Generative Predictive Control: Flow Matching Policies for Dynamic and Difficult-to-Demonstrate Tasks

no code implementations19 Feb 2025 Vince Kurtz, Joel W. Burdick

In particular, we introduce generative predictive control, a supervised learning framework for tasks with fast dynamics that are easy to simulate but difficult to demonstrate.

Minimizing Conservatism in Safety-Critical Control for Input-Delayed Systems via Adaptive Delay Estimation

no code implementations26 Nov 2024 Yitaek Kim, Ersin Das, Jeeseop Kim, Aaron D. Ames, Joel W. Burdick, Christoffer Sloth

The main idea is to reduce the maximum delay estimation error bound so that the state prediction error bound is monotonically non-increasing.

Prediction

Supervised Learning for Stochastic Optimal Control

no code implementations9 Sep 2024 Vince Kurtz, Joel W. Burdick

In this paper, we show how to automatically generate supervised learning data for a class of continuous-time nonlinear stochastic optimal control problems.

Rollover Prevention for Mobile Robots with Control Barrier Functions: Differentiator-Based Adaptation and Projection-to-State Safety

no code implementations13 Mar 2024 Ersin Das, Aaron D. Ames, Joel W. Burdick

This paper develops rollover prevention guarantees for mobile robots using control barrier function (CBF) theory, and demonstrates the method experimentally.

Robust Control Barrier Functions using Uncertainty Estimation with Application to Mobile Robots

no code implementations3 Jan 2024 Ersin Das, Joel W. Burdick

Our constructive framework couples control barrier function (CBF) theory with a new uncertainty estimator to ensure robust safety.

Learning Disturbances Online for Risk-Aware Control: Risk-Aware Flight with Less Than One Minute of Data

no code implementations12 Dec 2022 Prithvi Akella, Skylar X. Wei, Joel W. Burdick, Aaron D. Ames

Recent advances in safety-critical risk-aware control are predicated on apriori knowledge of the disturbances a system might face.

Sample-Based Bounds for Coherent Risk Measures: Applications to Policy Synthesis and Verification

no code implementations21 Apr 2022 Prithvi Akella, Anushri Dixit, Mohamadreza Ahmadi, Joel W. Burdick, Aaron D. Ames

The dramatic increase of autonomous systems subject to variable environments has given rise to the pressing need to consider risk in both the synthesis and verification of policies for these systems.

Risk-Averse Receding Horizon Motion Planning for Obstacle Avoidance using Coherent Risk Measures

no code implementations20 Apr 2022 Anushri Dixit, Mohamadreza Ahmadi, Joel W. Burdick

This paper studies the problem of risk-averse receding horizon motion planning for agents with uncertain dynamics, in the presence of stochastic, dynamic obstacles.

Model Predictive Control Motion Planning

Distributionally Robust Model Predictive Control with Total Variation Distance

no code implementations22 Mar 2022 Anushri Dixit, Mohamadreza Ahmadi, Joel W. Burdick

This paper studies the problem of distributionally robust model predictive control (MPC) using total variation distance ambiguity sets.

Computational Efficiency Model Predictive Control

Risk-Averse Stochastic Shortest Path Planning

no code implementations26 Mar 2021 Mohamadreza Ahmadi, Anushri Dixit, Joel W. Burdick, Aaron D. Ames

We consider the stochastic shortest path planning problem in MDPs, i. e., the problem of designing policies that ensure reaching a goal state from a given initial state with minimum accrued cost.

Learning Invariant Representation of Tasks for Robust Surgical State Estimation

no code implementations18 Feb 2021 Yidan Qin, Max Allan, Yisong Yue, Joel W. Burdick, Mahdi Azizian

The combination of high diversity and limited data calls for new learning methods that are robust and invariant to operating conditions and surgical techniques.

Diversity

Risk-Sensitive Motion Planning using Entropic Value-at-Risk

no code implementations23 Nov 2020 Anushri Dixit, Mohamadreza Ahmadi, Joel W. Burdick

We consider the problem of risk-sensitive motion planning in the presence of randomly moving obstacles.

Model Predictive Control Motion Planning

Safe Multi-Agent Interaction through Robust Control Barrier Functions with Learned Uncertainties

1 code implementation11 Apr 2020 Richard Cheng, Mohammad Javad Khojasteh, Aaron D. Ames, Joel W. Burdick

Robots operating in real world settings must navigate and maintain safety while interacting with many heterogeneous agents and obstacles.

Navigate

Human Preference-Based Learning for High-dimensional Optimization of Exoskeleton Walking Gaits

1 code implementation13 Mar 2020 Maegan Tucker, Myra Cheng, Ellen Novoseller, Richard Cheng, Yisong Yue, Joel W. Burdick, Aaron D. Ames

Optimizing lower-body exoskeleton walking gaits for user comfort requires understanding users' preferences over a high-dimensional gait parameter space.

Temporal Segmentation of Surgical Sub-tasks through Deep Learning with Multiple Data Sources

no code implementations7 Feb 2020 Yidan Qin, Sahba Aghajani Pedram, Seyedshams Feyzabadi, Max Allan, A. Jonathan McLeod, Joel W. Burdick, Mahdi Azizian

A crucial step towards the automation of such surgical tasks is the temporal perception of the current surgical scene, which requires a real-time estimation of the states in the FSMs.

Stochastic Finite State Control of POMDPs with LTL Specifications

no code implementations21 Jan 2020 Mohamadreza Ahmadi, Rangoli Sharan, Joel W. Burdick

Partially observable Markov decision processes (POMDPs) provide a modeling framework for autonomous decision making under uncertainty and imperfect sensing, e. g. robot manipulation and self-driving cars.

Decision Making Decision Making Under Uncertainty +3

Dueling Posterior Sampling for Preference-Based Reinforcement Learning

1 code implementation4 Aug 2019 Ellen R. Novoseller, Yibing Wei, Yanan Sui, Yisong Yue, Joel W. Burdick

In preference-based reinforcement learning (RL), an agent interacts with the environment while receiving preferences instead of absolute feedback.

reinforcement-learning Reinforcement Learning +1

Control Regularization for Reduced Variance Reinforcement Learning

1 code implementation14 May 2019 Richard Cheng, Abhinav Verma, Gabor Orosz, Swarat Chaudhuri, Yisong Yue, Joel W. Burdick

We show that functional regularization yields a bias-variance trade-off, and propose an adaptive tuning strategy to optimize this trade-off.

continuous-control Continuous Control +3

End-to-End Safe Reinforcement Learning through Barrier Functions for Safety-Critical Continuous Control Tasks

1 code implementation21 Mar 2019 Richard Cheng, Gabor Orosz, Richard M. Murray, Joel W. Burdick

Reinforcement Learning (RL) algorithms have found limited success beyond simulated applications, and one main reason is the absence of safety guarantees during the learning process.

continuous-control Continuous Control +4

Stagewise Safe Bayesian Optimization with Gaussian Processes

no code implementations ICML 2018 Yanan Sui, Vincent Zhuang, Joel W. Burdick, Yisong Yue

We provide theoretical guarantees for both the satisfaction of safety constraints as well as convergence to the optimal utility value.

Bayesian Optimization Decision Making +3

Quantifying Performance of Bipedal Standing with Multi-channel EMG

no code implementations21 Nov 2017 Yanan Sui, Kun Ho Kim, Joel W. Burdick

Spinal cord stimulation has enabled humans with motor complete spinal cord injury (SCI) to independently stand and recover some lost autonomic function.

Electromyography (EMG)

Meta Inverse Reinforcement Learning via Maximum Reward Sharing for Human Motion Analysis

no code implementations7 Oct 2017 Kun Li, Joel W. Burdick

Observing that each demonstrator has an inherent reward for each state and the task-specific behaviors mainly depend on a small number of key states, we propose a meta IRL algorithm that first models the reward function for each task as a distribution conditioned on a baseline reward function shared by all tasks and dependent only on the demonstrator, and then finds the most likely reward function in the distribution that explains the task-specific behaviors.

reinforcement-learning Reinforcement Learning +1

A Function Approximation Method for Model-based High-Dimensional Inverse Reinforcement Learning

no code implementations23 Aug 2017 Kun Li, Joel W. Burdick

This works handles the inverse reinforcement learning problem in high-dimensional state spaces, which relies on an efficient solution of model-based high-dimensional reinforcement learning problems.

reinforcement-learning Reinforcement Learning +1

Bellman Gradient Iteration for Inverse Reinforcement Learning

no code implementations24 Jul 2017 Kun Li, Yanan Sui, Joel W. Burdick

We introduce a strategy to flexibly handle different types of actions with two approximations of the Bellman Optimality Equation, and a Bellman Gradient Iteration method to compute the gradient of the Q-value with respect to the reward function.

reinforcement-learning Reinforcement Learning +1

Clinical Patient Tracking in the Presence of Transient and Permanent Occlusions via Geodesic Feature

no code implementations22 Jul 2017 Kun Li, Joel W. Burdick

This paper develops a method to use RGB-D cameras to track the motions of a human spinal cord injury patient undergoing spinal stimulation and physical rehabilitation.

Multi-dueling Bandits with Dependent Arms

no code implementations29 Apr 2017 Yanan Sui, Vincent Zhuang, Joel W. Burdick, Yisong Yue

The dueling bandits problem is an online learning framework for learning from pairwise preference feedback, and is particularly well-suited for modeling settings that elicit subjective or implicit human feedback.

Thompson Sampling

Convex Relaxations of SE(2) and SE(3) for Visual Pose Estimation

no code implementations15 Jan 2014 Matanya B. Horowitz, Nikolai Matni, Joel W. Burdick

The method is a convex relaxation of the classical pose estimation problem, and is based on explicit linear matrix inequality (LMI) representations for the convex hulls of $SE(2)$ and $SE(3)$.

Pose Estimation

Cannot find the paper you are looking for? You can Submit a new open access paper.