Search Results for author: Fengyuan Shi

Found 5 papers, 1 papers with code

BIVDiff: A Training-Free Framework for General-Purpose Video Synthesis via Bridging Image and Video Diffusion Models

1 code implementation5 Dec 2023 Fengyuan Shi, Jiaxi Gu, Hang Xu, Songcen Xu, Wei zhang, LiMin Wang

Now text-to-image foundation models are widely applied to various downstream image synthesis tasks, such as controllable image generation and image editing, while downstream video synthesis tasks are less explored for several reasons.

Image Generation Model Selection +3

Bridging The Gaps Between Token Pruning and Full Pre-training via Masked Fine-tuning

no code implementations26 Oct 2023 Fengyuan Shi, LiMin Wang

Despite the success of transformers on various computer vision tasks, they suffer from excessive memory and computational cost.

Progressive Visual Prompt Learning with Contrastive Feature Re-formation

no code implementations17 Apr 2023 Chen Xu, Haocheng Shen, Fengyuan Shi, Boheng Chen, Yixuan Liao, Xiaoxin Chen, LiMin Wang

To the best of our knowledge, we are the first to demonstrate the superior performance of visual prompts in V-L models to previous prompt-based methods in downstream tasks.

Dynamic MDETR: A Dynamic Multimodal Transformer Decoder for Visual Grounding

no code implementations28 Sep 2022 Fengyuan Shi, Ruopeng Gao, Weilin Huang, LiMin Wang

The sampling module aims to select these informative patches by predicting the offsets with respect to a reference point, while the decoding module works for extracting the grounded object information by performing cross attention between image features and text features.

Visual Grounding

End-to-End Dense Video Grounding via Parallel Regression

no code implementations23 Sep 2021 Fengyuan Shi, Weilin Huang, LiMin Wang

In this paper, we tackle a new problem of dense video grounding, by simultaneously localizing multiple moments with a paragraph as input.

regression Sentence +1

Cannot find the paper you are looking for? You can Submit a new open access paper.