Search Results for author: Sicheng Mo

Found 7 papers, 3 papers with code

SimGen: Simulator-conditioned Driving Scene Generation

no code implementations13 Jun 2024 Yunsong Zhou, Michael Simon, Zhenghao Peng, Sicheng Mo, Hongzi Zhu, Minyi Guo, Bolei Zhou

Moreover, the trained models can only generate images based on the real-world layout data from the validation set of the same dataset, where overfitting might happen.

Autonomous Driving Data Augmentation +3

Ctrl-X: Controlling Structure and Appearance for Text-To-Image Generation Without Guidance

no code implementations11 Jun 2024 Kuan Heng Lin, Sicheng Mo, Ben Klingher, Fangzhou Mu, Bolei Zhou

Ctrl-X designs feed-forward structure control to enable the structure alignment with a structure image and semantic-aware appearance transfer to facilitate the appearance transfer from a user-input image.

Text-to-Image Generation

SnAG: Scalable and Accurate Video Grounding

1 code implementation CVPR 2024 Fangzhou Mu, Sicheng Mo, Yin Li

In this paper, we study the effect of cross-modal fusion on the scalability of video grounding models.

Video Grounding Video Understanding

A Simple Transformer-Based Model for Ego4D Natural Language Queries Challenge

1 code implementation16 Nov 2022 Sicheng Mo, Fangzhou Mu, Yin Li

This report describes Badgers@UW-Madison, our submission to the Ego4D Natural Language Queries (NLQ) Challenge.

Natural Language Queries Temporal Action Localization +1

Physics to the Rescue: Deep Non-line-of-sight Reconstruction for High-speed Imaging

no code implementations3 May 2022 Fangzhou Mu, Sicheng Mo, Jiayong Peng, Xiaochun Liu, Ji Hyun Nam, Siddeshwar Raghavan, Andreas Velten, Yin Li

Computational approach to imaging around the corner, or non-line-of-sight (NLOS) imaging, is becoming a reality thanks to major advances in imaging hardware and reconstruction algorithms.

Cannot find the paper you are looking for? You can Submit a new open access paper.