Search Results for author: Prashant Doshi

Found 23 papers, 2 papers with code

MVSA-Net: Multi-View State-Action Recognition for Robust and Deployable Trajectory Generation

no code implementations14 Nov 2023 Ehsan Asali, Prashant Doshi, Jin Sun

Performance evaluations on two distinct domains establish that MVSA-Net recognizes the state-action pairs under occlusion more accurately compared to single-view MVSA-Net and other baselines.

Action Recognition

A Novel Variational Lower Bound for Inverse Reinforcement Learning

no code implementations7 Nov 2023 Yikang Gui, Prashant Doshi

In this paper, we present a new Variational Lower Bound for IRL (VLB-IRL), which is derived under the framework of a probabilistic graphical model with an optimality node.

reinforcement-learning valid

Latent Interactive A2C for Improved RL in Open Many-Agent Systems

no code implementations9 May 2023 Keyang He, Prashant Doshi, Bikramjit Banerjee

There is a prevalence of multiagent reinforcement learning (MARL) methods that engage in centralized training.

IRL with Partial Observations using the Principle of Uncertain Maximum Entropy

no code implementations15 Aug 2022 Kenneth Bogert, Yikang Gui, Prashant Doshi

We show that in generalizing the principle of maximum entropy to these types of scenarios we unavoidably introduce a dependency on the learned model to the empirical feature expectations.

Marginal MAP Estimation for Inverse RL under Occlusion with Observer Noise

no code implementations16 Sep 2021 Prasanth Sengadu Suresh, Prashant Doshi

We consider the problem of learning the behavioral preferences of an expert engaged in a task from noisy and partially-observable demonstrations.

A Hierarchical Bayesian model for Inverse RL in Partially-Controlled Environments

no code implementations13 Jul 2021 Kenneth Bogert, Prashant Doshi

To address this, we present a hierarchical Bayesian model that incorporates both the expert's and the confounding elements' observations thereby explicitly modeling the diverse observations a robot may receive.

Many Agent Reinforcement Learning Under Partial Observability

no code implementations17 Jun 2021 Keyang He, Prashant Doshi, Bikramjit Banerjee

Recent renewed interest in multi-agent reinforcement learning (MARL) has generated an impressive array of techniques that leverage deep reinforcement learning, primarily actor-critic architectures, and can be applied to a limited range of settings in terms of observability and communication.

Multi-agent Reinforcement Learning reinforcement-learning +1

Maximum Entropy Multi-Task Inverse RL

1 code implementation27 Apr 2020 Saurabh Arora, Bikramjit Banerjee, Prashant Doshi

The learner aims to learn the multiple reward functions that guide these ways of solving the problem.

Clustering

Online Structure Learning for Feed-Forward and Recurrent Sum-Product Networks

no code implementations NeurIPS 2018 Agastya Kalra, Abdullah Rashwan, Wei-Shou Hsu, Pascal Poupart, Prashant Doshi, Georgios Trimponias

Sum-product networks have recently emerged as an attractive representation due to their dual view as a special type of deep neural network with clear semantics and a special type of probabilistic graphical model for which inference is always tractable.

valid

A Survey of Inverse Reinforcement Learning: Challenges, Methods and Progress

no code implementations18 Jun 2018 Saurabh Arora, Prashant Doshi

By categorically surveying the current literature in IRL, this article serves as a reference for researchers and practitioners of machine learning and beyond to understand the challenges of IRL and select the approaches best suited for the problem on hand.

reinforcement-learning Reinforcement Learning (RL)

Reinforcement Learning for Heterogeneous Teams with PALO Bounds

no code implementations23 May 2018 Roi Ceren, Prashant Doshi, Keyang He

We introduce reinforcement learning for heterogeneous teams in which rewards for an agent are additively factored into local costs, stimuli unique to each agent, and global rewards, those shared by all agents in the domain.

reinforcement-learning Reinforcement Learning (RL)

A Framework and Method for Online Inverse Reinforcement Learning

no code implementations21 May 2018 Saurabh Arora, Prashant Doshi, Bikramjit Banerjee

Inverse reinforcement learning (IRL) is the problem of learning the preferences of an agent from the observations of its behavior on a task.

reinforcement-learning Reinforcement Learning (RL)

Actor-Critic for Linearly-Solvable Continuous MDP with Partially Known Dynamics

no code implementations4 Jun 2017 Tomoki Nishi, Prashant Doshi, Michael R. James, Danil Prokhorov

In many robotic applications, some aspects of the system dynamics can be modeled accurately while others are difficult to obtain or model.

Reinforcement Learning (RL)

Individual Planning in Infinite-Horizon Multiagent Settings: Inference, Structure and Scalability

no code implementations NeurIPS 2015 Xia Qu, Prashant Doshi

This paper provides the first formalization of self-interested planning in multiagent settings using expectation-maximization (EM).

Dynamic Sum Product Networks for Tractable Inference on Sequence Data (Extended Version)

no code implementations13 Nov 2015 Mazen Melibari, Pascal Poupart, Prashant Doshi, George Trimponias

Since SPNs represent distributions over a fixed set of variables only, we propose dynamic sum product networks (DSPNs) as a generalization of SPNs for sequence data of varying length.

Individual Planning in Agent Populations: Exploiting Anonymity and Frame-Action Hypergraphs

no code implementations24 Mar 2015 Ekhlas Sonu, Yingke Chen, Prashant Doshi

Traditional I-POMDPs model this dependence on the actions of other agents using joint action and model spaces.

Team Behavior in Interactive Dynamic Influence Diagrams with Applications to Ad Hoc Teams

no code implementations1 Sep 2014 Muthukumaran Chandrasekaran, Prashant Doshi, Yifeng Zeng, Yingke Chen

This motivates investigating the ad hoc teamwork problem in the context of individual decision making frameworks.

Decision Making

Exploiting Model Equivalences for Solving Interactive Dynamic Influence Diagrams

no code implementations18 Jan 2014 Yifeng Zeng, Prashant Doshi

I-DIDs are graphical models and may be used to compute the policy of an agent given its belief over the physical state and others models, which changes as the agent acts and observes in the multiagent setting.

Decision Making

Monte Carlo Sampling Methods for Approximating Interactive POMDPs

no code implementations15 Jan 2014 Prashant Doshi, Piotr J. Gmytrasiewicz

The interactive PF is able to mitigate the belief space complexity, but it does not address the policy space complexity.

Cannot find the paper you are looking for? You can Submit a new open access paper.