no code implementations • 28 Nov 2023 • Zixiang Zhou, Yu Wan, Baoyuan Wang
AvatarGPT treats each task as one type of instruction fine-tuned on the shared LLM.
no code implementations • 28 Nov 2023 • Zixiang Zhou, Yu Wan, Baoyuan Wang
The field has made significant progress in synthesizing realistic human motion driven by various modalities.
no code implementations • 23 Oct 2023 • Minkyoung Cho, Yulong Cao, Zixiang Zhou, Z. Morley Mao
Deep neural networks (DNNs) are increasingly integrated into LiDAR (Light Detection and Ranging)-based perception systems for autonomous vehicles (AVs), requiring robust performance under adversarial conditions.
1 code implementation • 4 Sep 2023 • Zixiang Zhou, Weiyuan Li, Baoyuan Wang
We found that directly measuring the embedding distance between motion and music is not an optimal solution.
1 code implementation • 21 Jul 2023 • Shengnan Hu, Ce Zheng, Zixiang Zhou, Chen Chen, Gita Sukthankar
Human-centric visual understanding is an important desideratum for effective human-robot interaction.
no code implementations • 21 Jun 2023 • Dongqiangzi Ye, Yufei Xie, Weijia Chen, Zixiang Zhou, Lingting Ge, Hassan Foroosh
Due to the difficulty of acquiring large-scale 3D human keypoint annotation, previous methods for 3D human pose estimation (HPE) have often relied on 2D image features and sequential 2D annotations.
no code implementations • 21 Mar 2023 • Zixiang Zhou, Dongqiangzi Ye, Weijia Chen, Yufei Xie, Yu Wang, Panqu Wang, Hassan Foroosh
The proposed LiDARFormer utilizes cross-space global contextual feature information and exploits cross-task synergy to boost the performance of LiDAR perception tasks across multiple large-scale datasets and benchmarks.
1 code implementation • CVPR 2023 • Zixiang Zhou, Baoyuan Wang
Generating controllable and editable human motion sequences is a key challenge in 3D Avatar generation.
no code implementations • 19 Sep 2022 • Dongqiangzi Ye, Zixiang Zhou, Weijia Chen, Yufei Xie, Yu Wang, Panqu Wang, Hassan Foroosh
LidarMultiNet is extensively tested on both Waymo Open Dataset and nuScenes dataset, demonstrating for the first time that major LiDAR perception tasks can be unified in a single strong network that is trained end-to-end and achieves state-of-the-art performance.
1 code implementation • 12 Sep 2022 • Zixiang Zhou, Xiangchen Zhao, Yu Wang, Panqu Wang, Hassan Foroosh
It then uses the feature of the center candidate as the query embedding in the transformer.
Ranked #2 on 3D Object Detection on waymo cyclist
no code implementations • 23 Jun 2022 • Dongqiangzi Ye, Weijia Chen, Zixiang Zhou, Yufei Xie, Yu Wang, Panqu Wang, Hassan Foroosh
This technical report presents the 1st place winning solution for the Waymo Open Dataset 3D semantic segmentation challenge 2022.
2 code implementations • CVPR 2021 • Zixiang Zhou, Yang Zhang, Hassan Foroosh
Panoptic segmentation presents a new challenge in exploiting the merits of both detection and segmentation, with the aim of unifying instance segmentation and semantic segmentation in a single framework.
4 code implementations • CVPR 2020 • Yang Zhang, Zixiang Zhou, Philip David, Xiangyu Yue, Zerong Xi, Boqing Gong, Hassan Foroosh
The need for fine-grained perception in autonomous driving systems has resulted in recently increased research on online semantic segmentation of single-scan LiDAR.
Ranked #11 on Robust 3D Semantic Segmentation on nuScenes-C