no code implementations • 30 Nov 2024 • Shipeng Liu, Boshen Zhang, Zhehui Huang
To address this gap, we propose a novel benchmark that assesses agents' reactive adaptability and instantaneous communication capabilities at every step.
no code implementations • 26 Nov 2024 • Shipeng Liu, FNU Shrutika, Boshen Zhang, Zhehui Huang, Feifei Qian
Leveraging the strong communication capabilities of Large Language Models (LLMs), we propose a Human-Robot Teaming Framework with Multi-Modal Language feedback (HRT-ML), a framework designed to enhance human-robot interaction by adjusting the frequency and content of language-based feedback.
1 code implementation • 12 Sep 2024 • Liqiang Jing, Zhehui Huang, Xiaoyang Wang, Wenlin Yao, Wenhao Yu, Kaixin Ma, Hongming Zhang, Xinya Du, Dong Yu
To bridge this gap, we introduce DSBench, a comprehensive benchmark designed to evaluate data science agents with realistic tasks.
1 code implementation • 16 Mar 2024 • Zhehui Huang, Guangyao Shi, Gaurav S. Sukhatme
We systematically investigate the performance of LLMs in robot routing by constructing a dataset with 80 unique robot routing problems across 8 variants in both single and multi-robot settings.
no code implementations • 8 Dec 2023 • K. R. Zentner, Ujjwal Puri, Zhehui Huang, Gaurav S. Sukhatme
Then, we show that introducing a "fixup" phase is sufficient to guarantee a trust region is enforced on every policy update while adding fewer than 5% additional gradient steps in practice.
no code implementations • 28 Sep 2023 • Shashank Hegde, Zhehui Huang, Gaurav S. Sukhatme
We demonstrate that the neural policies estimated by HyperPPO are capable of decentralized control of a Crazyflie2. 1 quadrotor.
no code implementations • 23 Sep 2023 • Zhehui Huang, Zhaojing Yang, Rahul Krupani, Baskın Şenbaşlar, Sumeet Batra, Gaurav S. Sukhatme
In this work, we propose an end-to-end DRL approach to control quadrotor swarms in environments with obstacles.
1 code implementation • 15 Jun 2023 • Zhehui Huang, Sumeet Batra, Tao Chen, Rahul Krupani, Tushar Kumar, Artem Molchanov, Aleksei Petrenko, James A. Preiss, Zhaojing Yang, Gaurav S. Sukhatme
In addition to speed, such simulators need to model the physics of the robots and their interaction with the environment to a level acceptable for transferring policies learned in simulation to reality.
4 code implementations • ICML 2020 • Aleksei Petrenko, Zhehui Huang, Tushar Kumar, Gaurav Sukhatme, Vladlen Koltun
In this work we aim to solve this problem by optimizing the efficiency and resource utilization of reinforcement learning algorithms instead of relying on distributed computation.