1 code implementation • 10 Dec 2024 • Jiayi Su, Youhe Feng, Zheng Li, Jinhua Song, Yangfan He, Botao Ren, Botian Xu
This paper presents a novel framework for modeling and conditional generation of 3D articulated objects.
no code implementations • 10 Oct 2024 • Feng Chen, Botian Xu, Pu Hua, Peiqi Duan, Yanchao Yang, Yi Ma, Huazhe Xu
For single-task quality, we evaluate the realism of the generated task and the completeness of the generated trajectories using large language models and vision-language models.
no code implementations • 24 Sep 2024 • Jiayu Chen, Chao Yu, Guosheng Li, Wenhao Tang, Xinyi Yang, Botian Xu, Huazhong Yang, Yu Wang
Multi-UAV pursuit-evasion, where pursuers aim to capture evaders, poses a key challenge for UAV swarm intelligence.
Deep Reinforcement Learning Multi-agent Reinforcement Learning
no code implementations • 5 Apr 2024 • Botao Ren, Botian Xu, Xue Yang, Yifan Pu, Jingyi Wang, Zhidong Deng
Additionally, we introduce self-supervised constraints on CLIP Tokens to ensure consistency.
no code implementations • 19 Dec 2023 • Jiayu Chen, Guosheng Li, Chao Yu, Xinyi Yang, Botian Xu, Huazhong Yang, Yu Wang
In this work, we introduce a dual curriculum learning framework, named DualCL, which addresses multi-UAV pursuit-evasion in diverse environments and demonstrates zero-shot transfer ability to unseen scenarios.
no code implementations • 28 Nov 2023 • Botao Ren, Botian Xu, Tengyu Liu, Jingyi Wang, Zhidong Deng
Neuroscience studies have shown that the human visual system utilizes high-level feedback information to guide lower-level perception, enabling adaptation to signals of different characteristics.
1 code implementation • 22 Sep 2023 • Botian Xu, Feng Gao, Chao Yu, Ruize Zhang, Yi Wu, Yu Wang
In this work, we introduce OmniDrones, an efficient and flexible platform tailored for reinforcement learning in drone control, built on Nvidia's Omniverse Isaac Sim.
1 code implementation • 3 Feb 2023 • Chao Yu, Jiaxuan Gao, Weilin Liu, Botian Xu, Hao Tang, Jiaqi Yang, Yu Wang, Yi Wu
A crucial limitation of this framework is that every policy in the pool is optimized w. r. t.
no code implementations • 21 Sep 2022 • Hui Bai, Ruimin Shen, Yue Lin, Botian Xu, Ran Cheng
In comparison with the state-of-the-art RLlib, we empirically demonstrate the unique advantages of Lamarckian on benchmark tests with up to 6000 CPU cores: i) both the sampling efficiency and training speed are doubled when running PPO on Google football game; ii) the training speed is 13 times faster when running PBT+PPO on Pong game.