Search Results for author: Fu-Yun Wang

Found 16 papers, 11 papers with code

GS-DiT: Advancing Video Generation with Pseudo 4D Gaussian Fields through Efficient Dense 3D Point Tracking

no code implementations5 Jan 2025 Weikang Bian, Zhaoyang Huang, Xiaoyu Shi, Yijin Li, Fu-Yun Wang, Hongsheng Li

Specifically, we propose a novel framework that constructs a pseudo 4D Gaussian field with dense 3D point tracking and renders the Gaussian field for all video frames.

Novel View Synthesis Point Tracking +1

BlinkVision: A Benchmark for Optical Flow, Scene Flow and Point Tracking Estimation using RGB Frames and Events

no code implementations27 Oct 2024 Yijin Li, Yichen Shen, Zhaoyang Huang, Shuo Chen, Weikang Bian, Xiaoyu Shi, Fu-Yun Wang, Keqiang Sun, Hujun Bao, Zhaopeng Cui, Guofeng Zhang, Hongsheng Li

BlinkVision enables extensive benchmarks on three types of correspondence tasks (optical flow, point tracking, and scene flow estimation) for both image-based and event-based methods, offering new observations, practices, and insights for future research.

Event-based vision Optical Flow Estimation +2

Stable Consistency Tuning: Understanding and Improving Consistency Models

1 code implementation24 Oct 2024 Fu-Yun Wang, Zhengyang Geng, Hongsheng Li

Diffusion models achieve superior generation quality but suffer from slow generation speed due to the iterative nature of denoising.

Denoising Image Generation

Trans4D: Realistic Geometry-Aware Transition for Compositional Text-to-4D Synthesis

no code implementations9 Oct 2024 Bohan Zeng, Ling Yang, Siyu Li, Jiaming Liu, Zixiang Zhang, Juanxi Tian, Kaixin Zhu, Yongzhen Guo, Fu-Yun Wang, Minkai Xu, Stefano Ermon, Wentao Zhang

Then we propose a geometry-aware 4D transition network to realize a complex scene-level 4D transition based on the plan, which involves expressive geometrical object deformation.

Video Generation

Rectified Diffusion: Straightness Is Not Your Need in Rectified Flow

1 code implementation9 Oct 2024 Fu-Yun Wang, Ling Yang, Zhaoyang Huang, Mengdi Wang, Hongsheng Li

Building on this insight, we propose Rectified Diffusion, which generalizes the design space and application scope of rectification to encompass the broader category of diffusion models, rather than being restricted to flow-matching models.

OSV: One Step is Enough for High-Quality Image to Video Generation

no code implementations17 Sep 2024 Xiaofeng Mao, Zhengkai Jiang, Fu-Yun Wang, Wenbing Zhu, Jiangning Zhang, Hao Chen, Mingmin Chi, Yabiao Wang

Video diffusion models have shown great potential in generating high-quality videos, making them an increasingly popular focus.

Image to Video Generation

Lumina-Next: Making Lumina-T2X Stronger and Faster with Next-DiT

1 code implementation5 Jun 2024 Le Zhuo, Ruoyi Du, Han Xiao, Yangguang Li, Dongyang Liu, Rongjie Huang, Wenze Liu, Lirui Zhao, Fu-Yun Wang, Zhanyu Ma, Xu Luo, Zehan Wang, Kaipeng Zhang, Xiangyang Zhu, Si Liu, Xiangyu Yue, Dingning Liu, Wanli Ouyang, Ziwei Liu, Yu Qiao, Hongsheng Li, Peng Gao

Lumina-T2X is a nascent family of Flow-based Large Diffusion Transformers that establishes a unified framework for transforming noise into various modalities, such as images and videos, conditioned on text instructions.

Point Cloud Generation Text-to-Image Generation

Rethinking the Spatial Inconsistency in Classifier-Free Diffusion Guidance

1 code implementation CVPR 2024 Dazhong Shen, Guanglu Song, Zeyue Xue, Fu-Yun Wang, Yu Liu

Classifier-Free Guidance (CFG) has been widely used in text-to-image diffusion models, where the CFG scale is introduced to control the strength of text guidance on the whole image space.

Denoising Semantic Segmentation

Be-Your-Outpainter: Mastering Video Outpainting through Input-Specific Adaptation

1 code implementation20 Mar 2024 Fu-Yun Wang, Xiaoshi Wu, Zhaoyang Huang, Xiaoyu Shi, Dazhong Shen, Guanglu Song, Yu Liu, Hongsheng Li

We introduce MOTIA Mastering Video Outpainting Through Input-Specific Adaptation, a diffusion-based pipeline that leverages both the intrinsic data-specific patterns of the source video and the image/video generative prior for effective outpainting.

AnimateLCM: Computation-Efficient Personalized Style Video Generation without Personalized Video Data

1 code implementation1 Feb 2024 Fu-Yun Wang, Zhaoyang Huang, Weikang Bian, Xiaoyu Shi, Keqiang Sun, Guanglu Song, Yu Liu, Hongsheng Li

This paper introduces an effective method for computation-efficient personalized style video generation without requiring access to any personalized video data.

Conditional Image Generation Denoising +2

Gen-L-Video: Multi-Text to Long Video Generation via Temporal Co-Denoising

1 code implementation29 May 2023 Fu-Yun Wang, Wenshuo Chen, Guanglu Song, Han-Jia Ye, Yu Liu, Hongsheng Li

To address this challenge, we introduce a novel paradigm dubbed as Gen-L-Video, capable of extending off-the-shelf short video diffusion models for generating and editing videos comprising hundreds of frames with diverse semantic segments without introducing additional training, all while preserving content consistency.

Denoising Image Generation +2

Forward Compatible Few-Shot Class-Incremental Learning

1 code implementation CVPR 2022 Da-Wei Zhou, Fu-Yun Wang, Han-Jia Ye, Liang Ma, ShiLiang Pu, De-Chuan Zhan

Forward compatibility requires future new classes to be easily incorporated into the current model based on the current stage data, and we seek to realize it by reserving embedding space for future new classes.

class-incremental learning Few-Shot Class-Incremental Learning +1

PyCIL: A Python Toolbox for Class-Incremental Learning

1 code implementation23 Dec 2021 Da-Wei Zhou, Fu-Yun Wang, Han-Jia Ye, De-Chuan Zhan

Traditional machine learning systems are deployed under the closed-world setting, which requires the entire training data before the offline training process.

BIG-bench Machine Learning class-incremental learning +2

Cannot find the paper you are looking for? You can Submit a new open access paper.