1 code implementation • 24 Oct 2021 • Sirui Chen, Yunhao Liu, Jialong Li, Shang Wen Yao, Tingxiang Fan, Jia Pan
We propose DiffSRL, a dynamic state representation learning pipeline utilizing differentiable simulation that can embed complex dynamics models as part of the end-to-end training.
no code implementations • The International Journal of Robotics Research 2020 • Tingxiang Fan, Pinxin Long, Wenxi Liu and Jia Pan
We validate the learned sensor-level collision-3avoidance policy in a variety of simulated and real-world scenarios with thorough performance evaluations for large-scale multi-robot systems.
2 code implementations • ECCV 2020 • Cheng Lin, Tingxiang Fan, Wenping Wang, Matthias Nießner
We explore how to enable machines to model 3D shapes like human modelers using deep reinforcement learning (RL).
no code implementations • 22 Oct 2019 • Tingxiang Fan, Pinxin Long, Wenxi Liu, Jia Pan, Ruigang Yang, Dinesh Manocha
Deep reinforcement learning has great potential to acquire complex, adaptive behaviors for autonomous agents automatically.
no code implementations • 4 Oct 2019 • Qingyang Tan, Tingxiang Fan, Jia Pan, Dinesh Manocha
We present a novel algorithm (DeepMNavigate) for global multi-agent navigation in dense scenarios using deep reinforcement learning (DRL).
no code implementations • 15 Nov 2018 • Fan Wang, Bo Zhou, Ke Chen, Tingxiang Fan, Xi Zhang, Jiangyong Li, Hao Tian, Jia Pan
We built neural networks as our policy to map sensor readings to control signals on the UAV.
no code implementations • 12 Sep 2018 • Zhe Hu, Jia Pan, Tingxiang Fan, Ruigang Yang, Dinesh Manocha
In this paper, we present a robotic navigation algorithm with natural language interfaces, which enables a robot to safely walk through a changing environment with moving persons by following human instructions such as "go to the restaurant and keep away from people".
2 code implementations • 28 Sep 2017 • Pinxin Long, Tingxiang Fan, Xinyi Liao, Wenxi Liu, Hao Zhang, Jia Pan
We validate the learned sensor-level collision avoidance policy in a variety of simulated scenarios with thorough performance evaluations and show that the final learned policy is able to find time efficient, collision-free paths for a large-scale robot system.