Search Results for author: Qinglin Lu

Found 7 papers, 3 papers with code

LoRA-Composer: Leveraging Low-Rank Adaptation for Multi-Concept Customization in Training-Free Diffusion Models

1 code implementation18 Mar 2024 Yang Yang, Wen Wang, Liang Peng, Chaotian Song, Yao Chen, Hengjia Li, Xiaolong Yang, Qinglin Lu, Deng Cai, Boxi Wu, Wei Liu

Customization generation techniques have significantly advanced the synthesis of specific concepts across varied contexts.

SmoothVideo: Smooth Video Synthesis with Noise Constraints on Diffusion Models for One-shot Video Tuning

1 code implementation29 Nov 2023 Liang Peng, Haoran Cheng, Zheng Yang, Ruisi Zhao, Linxuan Xia, Chaotian Song, Qinglin Lu, Boxi Wu, Wei Liu

By applying the loss to existing one-shot video tuning methods, we significantly improve the overall consistency and smoothness of the generated videos.

Multi-modal Segment Assemblage Network for Ad Video Editing with Importance-Coherence Reward

1 code implementation25 Sep 2022 Yunlong Tang, Siting Xu, Teng Wang, Qin Lin, Qinglin Lu, Feng Zheng

The existing method performs well at video segmentation stages but suffers from the problems of dependencies on extra cumbersome models and poor performance at the segment assemblage stage.

Video Editing Video Segmentation +1

Overview of Tencent Multi-modal Ads Video Understanding Challenge

no code implementations16 Sep 2021 Zhenzhi Wang, Liyu Wu, Zhimin Li, Jiangfeng Xiong, Qinglin Lu

Our challenge includes two tasks: video structuring in the temporal dimension and multi-modal video classification.

Multi-Label Classification Video Classification +1

Tencent AVS: A Holistic Ads Video Dataset for Multi-modal Scene Segmentation

no code implementations9 Dec 2022 Jie Jiang, Zhimin Li, Jiangfeng Xiong, Rongwei Quan, Qinglin Lu, Wei Liu

Therefore, TAVS is distinguished from previous temporal segmentation datasets due to its multi-modal information, holistic view of categories, and hierarchical granularities.

Multi-Label Classification Scene Segmentation +3

DialogGen: Multi-modal Interactive Dialogue System for Multi-turn Text-to-Image Generation

no code implementations13 Mar 2024 Minbin Huang, Yanxin Long, Xinchi Deng, Ruihang Chu, Jiangfeng Xiong, Xiaodan Liang, Hong Cheng, Qinglin Lu, Wei Liu

However, many of these works face challenges in identifying correct output modalities and generating coherent images accordingly as the number of output modalities increases and the conversations go deeper.

Prompt Engineering Text-to-Image Generation

Cannot find the paper you are looking for? You can Submit a new open access paper.