Search Results for author: Siddhant Bhambri

Found 9 papers, 1 papers with code

LLMs Can't Plan, But Can Help Planning in LLM-Modulo Frameworks

no code implementations2 Feb 2024 Subbarao Kambhampati, Karthik Valmeekam, Lin Guan, Kaya Stechly, Mudit Verma, Siddhant Bhambri, Lucas Saldyt, Anil Murthy

On the other side are perhaps over-pessimistic claims that all that LLMs are good for in planning/reasoning tasks are as mere translators of the problem specification from one syntactic format to another, and ship the problem off to external symbolic solvers.

Theory of Mind abilities of Large Language Models in Human-Robot Interaction : An Illusion?

no code implementations10 Jan 2024 Mudit Verma, Siddhant Bhambri, Subbarao Kambhampati

In this work, we explore the task of Perceived Behavior Recognition, where a robot employs a Large Language Model (LLM) to assess the robot's generated behavior in a manner similar to human observer.

Language Modelling Large Language Model

Benchmarking Multi-Agent Preference-based Reinforcement Learning for Human-AI Teaming

no code implementations21 Dec 2023 Siddhant Bhambri, Mudit Verma, Anil Murthy, Subbarao Kambhampati

We introduce the notion of Human-Flexibility, i. e. whether the human partner is amenable to multiple team strategies, with a special case being Specified Orchestration where the human has a single team policy in mind (most constrained case).

Benchmarking reinforcement-learning

Exploiting Unlabeled Data for Feedback Efficient Human Preference based Reinforcement Learning

no code implementations17 Feb 2023 Mudit Verma, Siddhant Bhambri, Subbarao Kambhampati

Preference Based Reinforcement Learning has shown much promise for utilizing human binary feedback on queried trajectory pairs to recover the underlying reward model of the Human in the Loop (HiL).

reinforcement-learning Reinforcement Learning (RL)

Reinforcement Learning Methods for Wordle: A POMDP/Adaptive Control Approach

no code implementations15 Nov 2022 Siddhant Bhambri, Amrita Bhattacharjee, Dimitri Bertsekas

In this paper we address the solution of the popular Wordle puzzle, using new reinforcement learning methods, which apply more generally to adaptive control of dynamic systems and to classes of Partially Observable Markov Decision Process (POMDP) problems.

reinforcement-learning Reinforcement Learning (RL)

Contrastively Learning Visual Attention as Affordance Cues from Demonstrations for Robotic Grasping

1 code implementation2 Apr 2021 Yantian Zha, Siddhant Bhambri, Lin Guan

In this work, our goal is instead to fill the gap between affordance discovery and affordance-based policy learning by integrating the two objectives in an end-to-end imitation learning framework based on deep neural networks.

Contrastive Learning Imitation Learning +1

Multi-objective Reinforcement Learning based approach for User-Centric Power Optimization in Smart Home Environments

no code implementations29 Sep 2020 Saurabh Gupta, Siddhant Bhambri, Karan Dhingra, Arun Balaji Buduru, Ponnurangam Kumaraguru

We experiment on real-world smart home data, and show that the multi-objective approaches: i) establish trade-off between the two objectives, ii) achieve better combined user satisfaction and power consumption than single-objective approaches.

Management Multi-Objective Reinforcement Learning

Making Smart Homes Smarter: Optimizing Energy Consumption with Human in the Loop

no code implementations6 Dec 2019 Mudit Verma, Siddhant Bhambri, Saurabh Gupta, Arun Balaji Buduru

Rapid advancements in the Internet of Things (IoT) have facilitated more efficient deployment of smart environment solutions for specific user requirement.

Clustering Reinforcement Learning (RL)

A Survey of Black-Box Adversarial Attacks on Computer Vision Models

no code implementations3 Dec 2019 Siddhant Bhambri, Sumanyu Muku, Avinash Tulasi, Arun Balaji Buduru

Machine learning has seen tremendous advances in the past few years, which has lead to deep learning models being deployed in varied applications of day-to-day life.

Cannot find the paper you are looking for? You can Submit a new open access paper.