Search Results for author: Guosheng Zhao

Found 5 papers, 0 papers with code

EgoVid-5M: A Large-Scale Video-Action Dataset for Egocentric Video Generation

no code implementations13 Nov 2024 XiaoFeng Wang, Kang Zhao, Feng Liu, Jiayu Wang, Guosheng Zhao, Xiaoyi Bao, Zheng Zhu, Yingya Zhang, Xingang Wang

Video generation has emerged as a promising tool for world simulation, leveraging visual data to replicate real-world environments.

Video Generation

DriveDreamer4D: World Models Are Effective Data Machines for 4D Driving Scene Representation

no code implementations17 Oct 2024 Guosheng Zhao, Chaojun Ni, XiaoFeng Wang, Zheng Zhu, Xueyang Zhang, Yida Wang, Guan Huang, Xinze Chen, Boyuan Wang, Youyi Zhang, Wenjun Mei, Xingang Wang

Contemporary sensor simulation methods, such as NeRF and 3DGS, rely predominantly on conditions closely aligned with training data distributions, which are largely confined to forward-driving scenarios.

4D reconstruction Autonomous Driving +1

CoReS: Orchestrating the Dance of Reasoning and Segmentation

no code implementations8 Apr 2024 Xiaoyi Bao, Siyang Sun, Shuailei Ma, Kecheng Zheng, Yuxin Guo, Guosheng Zhao, Yun Zheng, Xingang Wang

We believe that the act of reasoning segmentation should mirror the cognitive stages of human visual search, where each step is a progressive refinement of thought toward the final object.

Reasoning Segmentation Segmentation

DriveDreamer-2: LLM-Enhanced World Models for Diverse Driving Video Generation

no code implementations11 Mar 2024 Guosheng Zhao, XiaoFeng Wang, Zheng Zhu, Xinze Chen, Guan Huang, Xiaoyi Bao, Xingang Wang

DriveDreamer-2 is the first world model to generate customized driving videos, it can generate uncommon driving videos (e. g., vehicles abruptly cut in) in a user-friendly manner.

Autonomous Driving Language Modeling +3

Cannot find the paper you are looking for? You can Submit a new open access paper.