Search Results for author: Xinghang Li

Found 5 papers, 2 papers with code

Towards Generalist Robot Policies: What Matters in Building Vision-Language-Action Models

no code implementations18 Dec 2024 Xinghang Li, Peiyan Li, Minghuan Liu, Dong Wang, Jirong Liu, Bingyi Kang, Xiao Ma, Tao Kong, Hanbo Zhang, Huaping Liu

The obtained results convince us firmly to explain why we need VLA and develop a new family of VLAs, RoboVLMs, which require very few manual designs and achieve a new state-of-the-art performance in three simulation tasks and real-world experiments.

Ranked #5 on Robot Manipulation on SimplerEnv-Widow X (using extra training data)

Representation Learning Robot Manipulation +1

Sparse Spectral Training and Inference on Euclidean and Hyperbolic Neural Networks

no code implementations24 May 2024 Jialin Zhao, Yingtao Zhang, Xinghang Li, Huaping Liu, Carlo Vittorio Cannistraci

The growing computational demands posed by increasingly number of neural network's parameters necessitate low-memory-consumption training approaches.

Link Prediction Machine Translation +2

Unleashing Large-Scale Video Generative Pre-training for Visual Robot Manipulation

3 code implementations20 Dec 2023 Hongtao Wu, Ya Jing, Chilam Cheang, Guangzeng Chen, Jiafeng Xu, Xinghang Li, Minghuan Liu, Hang Li, Tao Kong

In this paper, we extend the scope of this effectiveness by showing that visual robot manipulation can significantly benefit from large-scale video generative pre-training.

Ranked #5 on Zero-shot Generalization on CALVIN (using extra training data)

Robot Manipulation Zero-shot Generalization

Vision-Language Foundation Models as Effective Robot Imitators

no code implementations2 Nov 2023 Xinghang Li, Minghuan Liu, Hanbo Zhang, Cunjun Yu, Jie Xu, Hongtao Wu, Chilam Cheang, Ya Jing, Weinan Zhang, Huaping Liu, Hang Li, Tao Kong

We believe RoboFlamingo has the potential to be a cost-effective and easy-to-use solution for robotics manipulation, empowering everyone with the ability to fine-tune their own robotics policy.

Imitation Learning

Mixed Neural Voxels for Fast Multi-view Video Synthesis

1 code implementation ICCV 2023 Feng Wang, Sinan Tan, Xinghang Li, Zeyue Tian, Yafei Song, Huaping Liu

In this paper, we present a novel method named MixVoxels to better represent the dynamic scenes with fast training speed and competitive rendering qualities.

Cannot find the paper you are looking for? You can Submit a new open access paper.