no code implementations • 5 Dec 2024 • James Queeney, Xiaoyi Cai, Mouhacine Benosman, Jonathan P. How
The reliable deployment of deep reinforcement learning in real-world settings requires the ability to generalize across a variety of conditions, including both in-distribution scenarios seen during training as well as novel out-of-distribution scenarios.
no code implementations • 30 Sep 2024 • Alan Papalia, Yulun Tian, David M. Rosen, Jonathan P. How, John J. Leonard
This paper presents an overview of the Burer-Monteiro method (BM), a technique that has been applied to solve robot perception problems to certifiable optimality in real-time.
no code implementations • 30 Sep 2024 • Nicholas Rober, Jonathan P. How
Refinement strategies such as partitioning or symbolic propagation are typically used to limit the conservativeness of RSOAs, but these approaches come with a high computational cost and often can only be used to verify safety for simple reachability problems.
no code implementations • 4 Sep 2024 • Xiaoyi Cai, James Queeney, Tong Xu, Aniket Datar, Chenhui Pan, Max Miller, Ashton Flather, Philip R. Osteen, Nicholas Roy, Xuesu Xiao, Jonathan P. How
Self-supervised learning is a powerful approach for developing traversability models for off-road navigation, but these models often struggle with inputs unseen during training.
no code implementations • 14 Jun 2024 • Kota Kondo, Claudius T. Tewari, Andrea Tagliabue, Jesus Tordesillas, Parker C. Lusk, Jonathan P. How
In decentralized multiagent trajectory planners, agents need to communicate and exchange their positions to generate collision-free trajectories.
no code implementations • 6 May 2024 • Minjae Cho, Jonathan P. How, Chuangchuang Sun
In this paper, we propose the MOOD-CRL (Model-based Offline OOD-Adapting Causal RL) algorithm, which aims to address the challenge of extrapolation for offline policy training through causal inference instead of policy-regularizing methods.
no code implementations • 2 May 2024 • Kota Kondo, Andrea Tagliabue, Xiaoyi Cai, Claudius Tewari, Olivia Garcia, Marcos Espitia-Alvarez, Jonathan P. How
Although the resulting NN policies are effective at quickly generating trajectories similar to those from the expert, (1) their output does not explicitly account for dynamic feasibility, and (2) the policies do not accommodate changes in the constraints different from those used during training.
no code implementations • 11 Feb 2024 • Parker C. Lusk, Jonathan P. How
When an informative initial estimation guess is available, the data association challenge is less acute; however, the existence of a high-quality initial guess is rare in most contexts.
no code implementations • 9 Jan 2024 • Annika Thomas, Jouko Kinnari, Parker Lusk, Kota Kondo, Jonathan P. How
SOS-Match is a promising new approach for landmark detection and correspondence search in unstructured environments that is robust to changes in lighting and appearance and is more computationally efficient than other approaches, suggesting that the geometric arrangement of segments is a valuable localization cue in unstructured environments.
Simultaneous Localization and Mapping Zero Shot Segmentation
no code implementations • 23 Nov 2023 • Andrea Tagliabue, Jonathan P. How
We tailor our approach to the task of localization and trajectory tracking on a multirotor, by learning a visuomotor policy that generates control actions using images from the onboard camera as only source of horizontal position.
2 code implementations • 10 Nov 2023 • Xiaoyi Cai, Siddharth Ancha, Lakshay Sharma, Philip R. Osteen, Bernadette Bucher, Stephen Phillips, Jiuguang Wang, Michael Everett, Nicholas Roy, Jonathan P. How
For uncertainty quantification, we efficiently model both aleatoric and epistemic uncertainty by learning discrete traction distributions and probability densities of the traction predictor's latent features.
no code implementations • 26 Oct 2023 • Jason J. Choi, Donggun Lee, Boyang Li, Jonathan P. How, Koushil Sreenath, Sylvia L. Herbert, Claire J. Tomlin
This strong link we establish between the reachability problem and the barrier constraint, while guaranteeing the continuity of the value function, is not achievable by previous backward reachability-based formulations.
1 code implementation • 17 Jul 2023 • Jesus Tordesillas, Jonathan P. How, Marco Hutter
This paper presents RAYEN, a framework to impose hard convex constraints on the output or latent variable of a neural network.
no code implementations • 29 Jun 2023 • Anthony Francis, Claudia Pérez-D'Arpino, Chengshu Li, Fei Xia, Alexandre Alahi, Rachid Alami, Aniket Bera, Abhijat Biswas, Joydeep Biswas, Rohan Chandra, Hao-Tien Lewis Chiang, Michael Everett, Sehoon Ha, Justin Hart, Jonathan P. How, Haresh Karnan, Tsang-Wei Edward Lee, Luis J. Manso, Reuth Mirksy, Sören Pirk, Phani Teja Singamaneni, Peter Stone, Ada V. Taylor, Peter Trautman, Nathan Tsoi, Marynel Vázquez, Xuesu Xiao, Peng Xu, Naoki Yokoyama, Alexander Toshev, Roberto Martín-Martín
A major challenge to deploying robots widely is navigation in human-populated environments, commonly referred to as social robot navigation.
no code implementations • 30 Mar 2023 • Evelyn Ruff, Rebecca Russell, Matthew Stoeckle, Piero Miotto, Jonathan P. How
This paper presents a novel methodology that uses surrogate models in the form of neural networks to reduce the computation time of simulation-based optimization of a reference trajectory.
no code implementations • 28 Mar 2023 • Tong Zhao, Andrea Tagliabue, Jonathan P. How
We tailor our approach to the task of learning an adaptive position and attitude control policy to track trajectories under challenging disturbances on a multirotor.
1 code implementation • 7 Mar 2023 • Stewart Jamieson, Jonathan P. How, Yogesh Girdhar
In our experiments, we show that DeepSeeColor offers comparable performance to the popular "Sea-Thru" algorithm (Akkaynak & Treibitz, 2019) while being able to rapidly process images at up to 60Hz, thus making it suitable for use onboard AUVs as a preprocessing step to enable more robust vision-based behaviours.
no code implementations • 24 Dec 2022 • Parker C. Lusk, Devarth Parikh, Jonathan P. How
Using geometric landmarks like lines and planes can increase navigation accuracy and decrease map storage requirements compared to commonly-used LiDAR point cloud maps.
no code implementations • 28 Oct 2022 • Dong-Ki Kim, Matthew Riemer, Miao Liu, Jakob N. Foerster, Gerald Tesauro, Jonathan P. How
By directly comparing active equilibria to Nash equilibria in these examples, we find that active equilibria find more effective solutions than Nash equilibria, concluding that an active equilibrium is the desired solution for multiagent learning settings.
no code implementations • 18 Oct 2022 • Andrea Tagliabue, Jonathan P. How
Thanks to the augmented data, we reduce the computation time and the number of demonstrations needed by IL, while providing robustness to sensing and process uncertainty.
no code implementations • 15 Oct 2022 • Parker C. Lusk, Kaveh Fathian, Jonathan P. How
We present a multiway fusion algorithm capable of directly processing uncertain pairwise affinities.
no code implementations • 14 Oct 2022 • Nicholas Rober, Michael Everett, Songan Zhang, Jonathan P. How
We introduce a hybrid partitioning method that uses both target set partitioning (TSP) and backreachable set partitioning (BRSP) to overcome a lower bound on estimation error that is present when using BRSP.
no code implementations • 28 Sep 2022 • Nicholas Rober, Sydney M. Katz, Chelsea Sidrane, Esen Yel, Michael Everett, Mykel J. Kochenderfer, Jonathan P. How
As neural networks (NNs) become more prevalent in safety-critical applications such as control of vehicles, there is a growing need to certify that systems with NN components are safe.
no code implementations • 28 Sep 2022 • Jacqueline Ankenbauer, Kaveh Fathian, Jonathan P. How
To demonstrate our framework, we consider an example of localizing a ground vehicle in a reference object map using only cars as objects.
no code implementations • 23 Sep 2022 • Lena M. Downes, Ted J. Steiner, Rebecca L. Russell, Jonathan P. How
This paper presents Restricted FOV Wide-Area Geolocalization (ReWAG), a cross-view geolocalization approach that generalizes WAG for use with standard, non-panoramic ground cameras by creating pose-aware embeddings and providing a strategy to incorporate particle pose into the Siamese network.
no code implementations • 20 Sep 2022 • Andrea Tagliabue, Yi-Hsuan Hsiao, Urban Fasel, J. Nathan Kutz, Steven L. Brunton, Yufeng Chen, Jonathan P. How
Accurate and agile trajectory tracking in sub-gram Micro Aerial Vehicles (MAVs) is challenging, as the small scale of the robot induces large model uncertainties, demanding robust feedback controllers, while the fast dynamics and computational constraints prevent the deployment of computationally expensive strategies.
2 code implementations • 14 Apr 2022 • Nicholas Rober, Michael Everett, Jonathan P. How
The increasing prevalence of neural networks (NNs) in safety-critical applications calls for methods to certify their behavior and guarantee safety.
1 code implementation • 25 Mar 2022 • Xiaoyi Cai, Michael Everett, Jonathan Fink, Jonathan P. How
Motion planning in off-road environments requires reasoning about both the geometry and semantics of the scene (e. g., a robot may be able to drive through soft bushes but not a fallen log).
no code implementations • 14 Mar 2022 • Macheng Shen, Jonathan P. How
Achieving the capability of adapting to ever-changing environments is a critical step towards building fully autonomous robots that operate safely in complicated scenarios.
no code implementations • 10 Mar 2022 • Lena M. Downes, Dong-Ki Kim, Ted J. Steiner, Jonathan P. How
Taken together, WAG's network training and particle filter weighting approach achieves city-scale position estimation accuracies on the order of 20 meters, a 98% reduction compared to a baseline training and weighting approach.
1 code implementation • 7 Mar 2022 • Dong-Ki Kim, Matthew Riemer, Miao Liu, Jakob N. Foerster, Michael Everett, Chuangchuang Sun, Gerald Tesauro, Jonathan P. How
An effective approach that has recently emerged for addressing this non-stationarity is for each agent to anticipate the learning of other agents and influence the evolution of future policies towards desirable behavior for its own benefit.
1 code implementation • 2 Oct 2021 • Qiangqiang Huang, Can Pu, Kasra Khosoussi, David M. Rosen, Dehann Fourie, Jonathan P. How, John J. Leonard
This paper presents normalizing flows for incremental smoothing and mapping (NF-iSAM), a novel algorithm for inferring the full posterior distribution in SLAM problems with nonlinear measurement models and non-Gaussian factors.
no code implementations • 21 Sep 2021 • Andrea Tagliabue, Dong-Ki Kim, Michael Everett, Jonathan P. How
Our approach opens the possibility of zero-shot transfer from a single demonstration collected in a nominal domain, such as a simulation or a robot in a lab/controlled environment, to a domain with bounded model errors/perturbations.
1 code implementation • 20 Sep 2021 • Marwa Abdulhai, Dong-Ki Kim, Matthew Riemer, Miao Liu, Gerald Tesauro, Jonathan P. How
Hierarchical reinforcement learning has focused on discovering temporally extended actions, such as options, that can provide benefits in problems requiring extensive exploration.
no code implementations • 14 Sep 2021 • Chuangchuang Sun, Dong-Ki Kim, Jonathan P. How
In a multirobot system, a number of cyber-physical attacks (e. g., communication hijack, observation perturbations) can challenge the robustness of agents.
1 code implementation • 9 Aug 2021 • Michael Everett, Golnaz Habibi, Chuangchuang Sun, Jonathan P. How
While the solutions are less tight than previous (semidefinite program-based) methods, they are substantially faster to compute, and some of those computational time savings can be used to refine the bounds through new input set partitioning techniques, which is shown to dramatically reduce the tightness gap.
1 code implementation • 28 Jun 2021 • Yulun Tian, Yun Chang, Fernando Herrera Arias, Carlos Nieto-Granda, Jonathan P. How, Luca Carlone
This paper presents Kimera-Multi, the first multi-robot system that (i) is robust and capable of identifying and rejecting incorrect inter and intra-robot loop closures resulting from perceptual aliasing, (ii) is fully distributed and only relies on local (peer-to-peer) communication to achieve distributed localization and mapping, and (iii) builds a globally consistent metric-semantic 3D mesh model of the environment in real-time, where faces of the mesh are annotated with semantic labels.
no code implementations • 25 Feb 2021 • Bruno Brito, Michael Everett, Jonathan P. How, Javier Alonso-Mora
Robotic navigation in environments shared with other robots or humans remains challenging because the intentions of the surrounding agents are not directly observable and the environment conditions are continuously changing.
1 code implementation • 5 Jan 2021 • Michael Everett, Golnaz Habibi, Jonathan P. How
Neural Networks (NNs) can provide major empirical performance improvements for robotic systems, but they also introduce challenges in formally analyzing those systems' safety properties.
2 code implementations • 20 Nov 2020 • Parker C. Lusk, Kaveh Fathian, Jonathan P. How
We formulate the problem in a graph-theoretic framework using the notion of geometric consistency.
no code implementations • 8 Nov 2020 • Yun Chang, Yulun Tian, Jonathan P. How, Luca Carlone
Our system, dubbed Kimera-Multi, is implemented by a team of robots equipped with visual-inertial sensors, and builds a 3D mesh model of the environment in real-time, where each face of the mesh is annotated with a semantic label (e. g., building, road, objects).
1 code implementation • 31 Oct 2020 • Dong-Ki Kim, Miao Liu, Matthew Riemer, Chuangchuang Sun, Marwa Abdulhai, Golnaz Habibi, Sebastian Lopez-Cot, Gerald Tesauro, Jonathan P. How
A fundamental challenge in multiagent reinforcement learning is to learn beneficial behaviors in a shared environment with other simultaneously learning agents.
no code implementations • 1 Oct 2020 • Michael Everett, Golnaz Habibi, Jonathan P. How
Recent works approximate the propagation of sets through nonlinear activations or partition the uncertainty set to provide a guaranteed outer bound on the set of possible NN outputs.
no code implementations • 15 Jul 2020 • Lena M. Downes, Ted J. Steiner, Jonathan P. How
Terrain relative navigation can improve the precision of a spacecraft's position estimate by detecting global features that act as supplementary measurements to correct for drift in the inertial navigation system.
no code implementations • 19 Jun 2020 • Chuangchuang Sun, Dong-Ki Kim, Jonathan P. How
To drive the constraint violation monotonically decrease, we take the constraints as Lyapunov functions and impose new linear constraints on the policy parameters' updating dynamics.
no code implementations • 11 Apr 2020 • Michael Everett, Bjorn Lutjens, Jonathan P. How
Deep Neural Network-based systems are now the state-of-the-art in many robotics tasks, but their application in safety-critical domains remains dangerous without formal guarantees on network robustness.
no code implementations • 10 Mar 2020 • Stewart Jamieson, Jonathan P. How, Yogesh Girdhar
We present a novel POMDP problem formulation for a robot that must autonomously decide where to go to collect new and scientifically relevant images given a limited ability to communicate with its human operator.
1 code implementation • 4 Mar 2020 • Parker C. Lusk, Xiaoyi Cai, Samir Wadhwania, Aleix Paris, Kaveh Fathian, Jonathan P. How
While solutions using onboard localization address the dependency on external infrastructure, the associated coordination strategies typically lack collision avoidance and scalability.
Robotics Multiagent Systems
no code implementations • 4 Mar 2020 • Andrea Tagliabue, Aleix Paris, Suhan Kim, Regan Kubicek, Sarah Bergbreiter, Jonathan P. How
Disturbance estimation for Micro Aerial Vehicles (MAVs) is crucial for robustness and safety.
no code implementations • 2 Mar 2020 • Chuangchuang Sun, Macheng Shen, Jonathan P. How
Through this sparsity structure, the agents can communicate in an effective as well as efficient way via only selectively attending to agents that matter the most and thus the scale of the MARL problem is reduced with little optimality compromised.
1 code implementation • 16 Feb 2020 • Rose E. Wang, Michael Everett, Jonathan P. How
There are several real-world tasks that would benefit from applying multiagent reinforcement learning (MARL) algorithms, including the coordination among self-driving cars.
no code implementations • 18 Jan 2020 • Samaneh Hosseini Semnani, Hugh Liu, Michael Everett, Anton de Ruiter, Jonathan P. How
This paper introduces a hybrid algorithm of deep reinforcement learning (RL) and Force-based motion planning (FMP) to solve distributed motion planning problem in dense and dynamic environments.
2 code implementations • 9 Jan 2020 • Jesus Tordesillas, Brett T. Lopez, Michael Everett, Jonathan P. How
The standard approaches that ensure safety by enforcing a "stop" condition in the free-known space can severely limit the speed of the vehicle, especially in situations where much of the world is unknown.
no code implementations • 21 Nov 2019 • Golnaz Habibi, Nikita Japuria, Jonathan P. How
This paper presents a novel incremental learning algorithm for pedestrian motion prediction, with the ability to improve the learned model over time when data is incrementally available.
no code implementations • 28 Oct 2019 • Björn Lütjens, Michael Everett, Jonathan P. How
Deep Neural Network-based systems are now the state-of-the-art in many robotics tasks, but their application in safety-critical domains remains dangerous without formal guarantees on network robustness.
3 code implementations • 24 Oct 2019 • Michael Everett, Yu Fan Chen, Jonathan P. How
Collision avoidance algorithms are essential for safe and efficient robot operation among pedestrians.
no code implementations • 18 Sep 2019 • Macheng Shen, Jonathan P. How
In order to achieve a good trade-off between the robustness of the learned policy and the computation complexity, we propose to train a separate opponent policy against the protagonist agent for evaluation purposes.
1 code implementation • 11 Sep 2019 • Arpan Kusari, Jonathan P. How
A Gaussian process is used to obtain a smooth interpolation over the reward function weights of the optimal value function for three well-known examples: GridWorld, Objectworld and Pendulum.
Autonomous Vehicles Multi-Objective Reinforcement Learning +1
1 code implementation • 24 Aug 2019 • Michael Everett, Justin Miller, Jonathan P. How
Context is key information about structured environments that could guide exploration toward the unknown goal location, but the abstract idea is difficult to quantify for use in a planning algorithm.
no code implementations • 15 Mar 2019 • Samir Wadhwania, Dong-Ki Kim, Shayegan Omidshafiei, Jonathan P. How
Multiagent reinforcement learning algorithms (MARL) have been demonstrated on complex tasks that require the coordination of a team of multiple agents to complete.
3 code implementations • 8 Mar 2019 • Jesus Tordesillas, Brett T. Lopez, Jonathan P. How
The desire of maintaining computational tractability typically leads to optimization problems that do not include the obstacles (collision checks are done on the solutions) or to formulations that use a convex decomposition of the free space and then impose an ad hoc allocation of each interval of the trajectory in a specific polyhedron.
Robotics
no code implementations • 7 Mar 2019 • Dong-Ki Kim, Miao Liu, Shayegan Omidshafiei, Sebastian Lopez-Cot, Matthew Riemer, Golnaz Habibi, Gerald Tesauro, Sami Mourad, Murray Campbell, Jonathan P. How
Collective learning can be greatly enhanced when agents effectively exchange knowledge with their peers.
no code implementations • 2 Mar 2019 • Yulun Tian, Kasra Khosoussi, Jonathan P. How
The so-called Burer-Monteiro method is a well-studied technique for solving large-scale semidefinite programs (SDPs) via low-rank factorization.
no code implementations • 14 Feb 2019 • Macheng Shen, Jonathan P. How
We pose an active perception problem where an autonomous agent actively interacts with a second agent with potentially adversarial behaviors.
1 code implementation • 6 Feb 2019 • Kaveh Fathian, Kasra Khosoussi, Yulun Tian, Parker Lusk, Jonathan P. How
Many robotics applications require alignment and fusion of observations obtained at multiple views to form a global model of the environment.
no code implementations • 19 Oct 2018 • Björn Lütjens, Michael Everett, Jonathan P. How
The importance of predictions that are robust to this distributional shift is evident for safety-critical applications, such as collision avoidance around pedestrians.
2 code implementations • 2 Oct 2018 • Jesus Tordesillas, Brett T. Lopez, John Carter, John Ware, Jonathan P. How
However, in unknown environments, this approach can lead to erratic or unstable behavior due to the interaction between the global planner, whose solution is changing constantly, and the local planner; a consequence of not capturing higher-order dynamics in the global plan.
Robotics
no code implementations • 25 Sep 2018 • Kristoffer M. Frey, Ted J. Steiner, Jonathan P. How
Data association in SLAM is fundamentally challenging, and handling ambiguity well is crucial to achieve robust operation in real-world environments.
Robotics
no code implementations • 25 Jun 2018 • Golnaz Habibi, Nikita Jaipuria, Jonathan P. How
This paper presents a novel context-based approach for pedestrian motion prediction in crowded, urban intersections, with the additional flexibility of prediction in similar, but new, environments.
no code implementations • 25 Jun 2018 • Nikita Jaipuria, Golnaz Habibi, Jonathan P. How
This paper presents a novel framework for accurate pedestrian intent prediction at intersections.
no code implementations • 20 May 2018 • Shayegan Omidshafiei, Dong-Ki Kim, Miao Liu, Gerald Tesauro, Matthew Riemer, Christopher Amato, Murray Campbell, Jonathan P. How
The problem of teaching to improve agent learning has been investigated by prior works, but these approaches make assumptions that prevent application of teaching to general multiagent problems, or require domain expertise for problems they can apply to.
6 code implementations • 4 May 2018 • Michael Everett, Yu Fan Chen, Jonathan P. How
This work extends our previous approach to develop an algorithm that learns collision avoidance among a variety of types of dynamic agents without assuming they follow any particular behavior rules.
no code implementations • 15 Mar 2018 • Macheng Shen, Golnaz Habibi, Jonathan P. How
We are interested in developing transfer learning algorithms that can be trained on the pedestrian trajectories collected at one intersection and yet still provide accurate predictions of the trajectories at another, previously unseen intersection.
1 code implementation • 28 Nov 2017 • Shayegan Omidshafiei, Dong-Ki Kim, Jason Pazis, Jonathan P. How
This paper presents the Crossmodal Attentive Skill Learner (CASL), integrated with the recently-introduced Asynchronous Advantage Option-Critic (A2OC) architecture [Harb et al., 2017] to enable hierarchical reinforcement learning across multiple sensory inputs.
no code implementations • 20 Sep 2017 • Kristoffer M. Frey, Ted J. Steiner, Jonathan P. How
To demonstrate why, we propose a quantitative metric called elimination complexity (EC) that bridges the existing analytic gap between graph structure and computation.
Robotics
no code implementations • 24 Jul 2017 • Miao Liu, Kavinayan Sivakumar, Shayegan Omidshafiei, Christopher Amato, Jonathan P. How
We implement two variants of multi-robot Search and Rescue (SAR) domains (with and without obstacles) on hardware to demonstrate the learned policies can effectively control a team of distributed robots to cooperate in a partially observable stochastic environment.
2 code implementations • 26 Mar 2017 • Yu Fan Chen, Michael Everett, Miao Liu, Jonathan P. How
For robotic vehicles to navigate safely and efficiently in pedestrian-rich environments, it is important to model subtle human behaviors and navigation rules (e. g., passing on the right).
no code implementations • ICML 2017 • Shayegan Omidshafiei, Jason Pazis, Christopher Amato, Jonathan P. How, John Vian
Many real-world tasks involve multiple agents with partial observability and limited communication.
Multi-agent Reinforcement Learning reinforcement-learning +2
no code implementations • NeurIPS 2016 • Jason Pazis, Ronald E. Parr, Jonathan P. How
We present the first application of the median of means in a PAC exploration algorithm for MDPs.
no code implementations • 26 Sep 2016 • Yu Fan Chen, Miao Liu, Michael Everett, Jonathan P. How
Finding feasible, collision-free paths for multiagent systems can be challenging, particularly in non-communicating scenarios where each agent's intent (e. g. goal) is unobservable to the others.
Multiagent Systems
no code implementations • CVPR 2015 • Julian Straub, Trevor Campbell, Jonathan P. How, John W. Fisher III
Based on the small-variance limit of Bayesian nonparametric von-Mises-Fisher (vMF) mixture distributions, we propose two new flexible and efficient k-means-like clustering algorithms for directional data such as surface normals.
no code implementations • 3 May 2016 • Shayegan Omidshafiei, Brett T. Lopez, Jonathan P. How, John Vian
This paper presents an approach for filtering sequences of object classification probabilities using online modeling of the noise characteristics of the classifier outputs.
no code implementations • CVPR 2017 • Julian Straub, Trevor Campbell, Jonathan P. How, John W. Fisher III
Point cloud alignment is a common problem in computer vision and robotics, with applications ranging from 3D object recognition to reconstruction.
no code implementations • NeurIPS 2015 • Trevor Campbell, Julian Straub, John W. Fisher III, Jonathan P. How
This paper presents a methodology for creating streaming, distributed inference algorithms for Bayesian nonparametric (BNP) models.
no code implementations • 1 May 2015 • Miao Liu, Christopher Amato, Xuejun Liao, Lawrence Carin, Jonathan P. How
Expectation maximization (EM) has recently been shown to be an efficient algorithm for learning finite-state controllers (FSCs) in large decentralized POMDPs (Dec-POMDPs).
no code implementations • 20 Feb 2015 • Shayegan Omidshafiei, Ali-akbar Agha-mohammadi, Christopher Amato, Jonathan P. How
To allow for a high-level representation that is natural for multi-robot problems and scalable to large discrete and continuous problems, this paper extends the Dec-POMDP model to the decentralized partially observable semi-Markov decision process (Dec-POSMDP).
no code implementations • 22 May 2014 • Sarah Ferguson, Brandon Luders, Robert C. Grande, Jonathan P. How
To plan safe trajectories in urban environments, autonomous vehicles must be able to quickly assess the future intentions of dynamic agents.
Robotics 68T40
no code implementations • 28 Mar 2014 • Trevor Campbell, Jonathan P. How
The method first employs variational inference on each individual learning agent to generate a local approximate posterior, the agents transmit their local posteriors to other agents in the network, and finally each agent combines its set of received local posteriors.
no code implementations • 12 Feb 2014 • Christopher Amato, George D. Konidaris, Gabriel Cruz, Christopher A. Maynor, Jonathan P. How, Leslie P. Kaelbling
We describe a probabilistic framework for synthesizing control policies for general multi-robot systems, given environment and sensor models and a cost function.
no code implementations • NeurIPS 2013 • Daniel S. Levine, Jonathan P. How
We consider the sensor selection problem on multivariate Gaussian distributions where only a \emph{subset} of latent variables is of inferential interest.
1 code implementation • NeurIPS 2013 • Trevor Campbell, Miao Liu, Brian Kulis, Jonathan P. How, Lawrence Carin
This paper presents a novel algorithm, based upon the dependent Dirichlet process mixture model (DDPMM), for clustering batch-sequential data containing an unknown number of evolving clusters.