1 code implementation • 10 Dec 2024 • Yingying Deng, Xiangyu He, Changwang Mei, Peisong Wang, Fan Tang
Though Rectified Flows (ReFlows) with distillation offers a promising way for fast sampling, its fast inversion transforms images back to structured noise for recovery and following editing remains unsolved.
Ranked #9 on Text-based Image Editing on PIE-Bench
no code implementations • 28 Nov 2024 • Yingying Deng, Xiangyu He, Fan Tang, WeiMing Dong
In contrast to existing approaches, we have discovered that latent features in vanilla diffusion models inherently contain natural style and content distributions.
no code implementations • 26 Nov 2024 • Ziyi Xu, Ziyao Huang, Juan Cao, Yong Zhang, Xiaodong Cun, Qing Shuai, Yuchen Wang, Linchao Bao, Jintao Li, Fan Tang
The automatic generation of anchor-style product promotion videos presents promising opportunities in online commerce, advertising, and consumer engagement.
no code implementations • 23 Nov 2024 • Xiaoyue Mi, Fan Tang, Juan Cao, Qiang Sheng, Ziyao Huang, Peng Li, Yang Liu, Tong-Yee Lee
To address these limitations, we propose DyEval, an LLM-powered dynamic interactive visual assessment framework that facilitates collaborative evaluation between humans and generative models for text-to-image systems.
no code implementations • 22 Nov 2024 • Yu Xu, Fan Tang, Juan Cao, Yuxin Zhang, Xiaoyu Kong, Jintao Li, Oliver Deussen, Tong-Yee Lee
Diffusion Transformers (DiTs) have exhibited robust capabilities in image generation tasks.
1 code implementation • 16 Aug 2024 • Xiaoyue Mi, Fan Tang, Juan Cao, Peng Li, Yang Liu
Qualitative and quantitative experiments validate that VCPro achieves a better trade-off between the visibility of perturbations and protection effectiveness, effectively prioritizing the protection of target concepts in images with less perceptible perturbations.
no code implementations • 28 Apr 2024 • Yunbing Jia, Xiaoyu Kong, Fan Tang, Yixing Gao, WeiMing Dong, Yi Yang
In this paper, we reveal the two sides of data augmentation: enhancements in closed-set recognition correlate with a significant decrease in open-set recognition.
1 code implementation • CVPR 2024 • You Wu, Kean Liu, Xiaoyue Mi, Fan Tang, Juan Cao, Jintao Li
Extensive experiments on various kinds of visual attributes with SOTA personalization methods show the ability of the proposed method to mimic target visual appearance in novel contexts, thus improving the controllability and flexibility of personalization.
no code implementations • 28 Mar 2024 • Yu Xu, Fan Tang, Juan Cao, Yuxin Zhang, Oliver Deussen, WeiMing Dong, Jintao Li, Tong-Yee Lee
Based on the adapters broken apart for separate training content and style, we then make the entity parameter space by reconstructing the content and style PLPs matrices, followed by fine-tuning the combined adapter to generate the target object with the desired appearance.
1 code implementation • CVPR 2024 • Ziyao Huang, Fan Tang, Yong Zhang, Xiaodong Cun, Juan Cao, Jintao Li, Tong-Yee Lee
We adopt a two-stage training strategy for the diffusion model, effectively binding movements with specific appearances.
1 code implementation • 25 Jan 2024 • Nisha Huang, WeiMing Dong, Yuxin Zhang, Fan Tang, Ronghui Li, Chongyang Ma, Xiu Li, Changsheng Xu
Large-scale text-to-image generative models have made impressive strides, showcasing their ability to synthesize a vast array of high-quality images.
1 code implementation • CVPR 2024 • Yingying Deng, Xiangyu He, Fan Tang, WeiMing Dong
Despite the remarkable progress in image style transfer formulating style in the context of art is inherently subjective and challenging.
1 code implementation • 8 Dec 2023 • Yuxin Zhang, Fan Tang, Nisha Huang, Haibin Huang, Chongyang Ma, WeiMing Dong, Changsheng Xu
The essence of a video lies in its dynamic motions, including character actions, object movements, and camera movements.
1 code implementation • 29 Nov 2023 • Xiaoyue Mi, Fan Tang, Zonghan Yang, Danding Wang, Juan Cao, Peng Li, Yang Liu
Despite the remarkable advances that have been made in continual learning, the adversarial vulnerability of such methods has not been fully discussed.
1 code implementation • 29 Nov 2023 • Xiaoyue Mi, Fan Tang, Yepeng Weng, Danding Wang, Juan Cao, Sheng Tang, Peng Li, Yang Liu
Despite the effectiveness in improving the robustness of neural networks, adversarial training has suffered from the natural accuracy degradation problem, i. e., accuracy on natural samples has reduced significantly.
1 code implementation • 25 Nov 2023 • Yingying Deng, Xiangyu He, Fan Tang, WeiMing Dong
Despite the remarkable progress in image style transfer, formulating style in the context of art is inherently subjective and challenging.
no code implementations • 20 Oct 2023 • Haipeng Fang, Zhihao Sun, Ziyao Huang, Fan Tang, Juan Cao, Sheng Tang
The advancement of generative AI has extended to the realm of Human Dance Generation, demonstrating superior generative capacities.
3 code implementations • 25 May 2023 • Yuxin Zhang, WeiMing Dong, Fan Tang, Nisha Huang, Haibin Huang, Chongyang Ma, Tong-Yee Lee, Oliver Deussen, Changsheng Xu
We apply ProSpect in various personalized attribute-aware image generation applications, such as image-guided or text-driven manipulations of materials, style, and layout, achieving previously unattainable results from a single image input without fine-tuning the diffusion models.
1 code implementation • CVPR 2023 • Tianyun Yang, Danding Wang, Fan Tang, Xinying Zhao, Juan Cao, Sheng Tang
In this study, we focus on a challenging task, namely Open-Set Model Attribution (OSMA), to simultaneously attribute images to known models and identify those from unknown ones.
1 code implementation • 9 Mar 2023 • Yuxin Zhang, Fan Tang, WeiMing Dong, Haibin Huang, Chongyang Ma, Tong-Yee Lee, Changsheng Xu
Our framework consists of three key components, i. e., a parallel contrastive learning scheme for style representation and style transfer, a domain enhancement module for effective learning of style distribution, and a generative network for style transfer.
1 code implementation • 23 Feb 2023 • Nisha Huang, Fan Tang, WeiMing Dong, Tong-Yee Lee, Changsheng Xu
Different from current mask-based image editing methods, we propose a novel region-aware diffusion model (RDM) for entity-level image editing, which could automatically locate the region of interest and replace it following given text prompts.
1 code implementation • CVPR 2023 • Yuxin Zhang, Nisha Huang, Fan Tang, Haibin Huang, Chongyang Ma, WeiMing Dong, Changsheng Xu
Our key idea is to learn artistic style directly from a single painting and then guide the synthesis without providing complex textual descriptions.
Ranked #5 on Style Transfer on StyleBench
1 code implementation • 19 Nov 2022 • Nisha Huang, Yuxin Zhang, Fan Tang, Chongyang Ma, Haibin Huang, Yong Zhang, WeiMing Dong, Changsheng Xu
Despite the impressive results of arbitrary image-guided style transfer methods, text-driven image stylization has recently been proposed for transferring a natural image into a stylized one according to textual descriptions of the target style provided by the user.
1 code implementation • 27 Sep 2022 • Nisha Huang, Fan Tang, WeiMing Dong, Changsheng Xu
Extensive experimental results on the quality and quantity of the generated digital art paintings confirm the effectiveness of the combination of the diffusion model and multimodal guidance.
1 code implementation • CVPR 2023 • Dihe Huang, Ying Chen, Shang Xu, Yong liu, Wenlong Wu, Yikang Ding, Chengjie Wang, Fan Tang
The detector-free feature matching approaches are currently attracting great attention thanks to their excellent performance.
1 code implementation • 19 May 2022 • Yuxin Zhang, Fan Tang, WeiMing Dong, Haibin Huang, Chongyang Ma, Tong-Yee Lee, Changsheng Xu
Our framework consists of three key components, i. e., a multi-layer style projector for style code encoding, a domain enhancement module for effective learning of style distribution, and a generative network for image style transfer.
Ranked #4 on Style Transfer on StyleBench
1 code implementation • CVPR 2022 • Hanjun Li, Xingjia Pan, Ke Yan, Fan Tang, Wei-Shi Zheng
Object detection under imperfect data receives great attention recently.
1 code implementation • 26 Jan 2022 • Chengcheng Ma, Xingjia Pan, Qixiang Ye, Fan Tang, WeiMing Dong, Changsheng Xu
Semi-supervised object detection has recently achieved substantial progress.
3 code implementations • CVPR 2022 • Yingying Deng, Fan Tang, WeiMing Dong, Chongyang Ma, Xingjia Pan, Lei Wang, Changsheng Xu
The goal of image style transfer is to render an image with artistic features guided by a style reference while maintaining the original content.
Ranked #3 on Style Transfer on StyleBench
1 code implementation • ICCV 2021 • Shulan Ruan, Yong Zhang, Kun Zhang, Yanbo Fan, Fan Tang, Qi Liu, Enhong Chen
Text-to-image synthesis refers to generating an image from a given text description, the key goal of which lies in photo realism and semantic consistency.
4 code implementations • 30 May 2021 • Yingying Deng, Fan Tang, WeiMing Dong, Chongyang Ma, Xingjia Pan, Lei Wang, Changsheng Xu
The goal of image style transfer is to render an image with artistic features guided by a style reference while maintaining the original content.
1 code implementation • CVPR 2021 • Xingjia Pan, Yingguo Gao, Zhiwen Lin, Fan Tang, WeiMing Dong, Haolei Yuan, Feiyue Huang, Changsheng Xu
Weakly supervised object localization(WSOL) remains an open problem given the deficiency of finding object extent information using a classification network.
no code implementations • 17 Sep 2020 • Yingying Deng, Fan Tang, Wei-Ming Dong, Haibin Huang, Chongyang Ma, Changsheng Xu
Towards this end, we propose Multi-Channel Correction network (MCCNet), which can be trained to fuse the exemplar style features and input content features for efficient style transfer while naturally maintaining the coherence of input videos.
no code implementations • 2 Jun 2020 • Minxuan Lin, Fan Tang, Wei-Ming Dong, Xiao Li, Chongyang Ma, Changsheng Xu
Currently, there are few methods that can perform both multimodal and multi-domain stylization simultaneously.
2 code implementations • 27 May 2020 • Yingying Deng, Fan Tang, Wei-Ming Dong, Wen Sun, Feiyue Huang, Changsheng Xu
Arbitrary style transfer is a significant topic with research value and application prospect.
no code implementations • 26 Feb 2020 • Minxuan Lin, Yingying Deng, Fan Tang, Wei-Ming Dong, Changsheng Xu
Controllable painting generation plays a pivotal role in image stylization.
no code implementations • 12 Feb 2018 • Fan Tang, Wei-Ming Dong, Yiping Meng, Chongyang Ma, Fuzhang Wu, Xinrui Li, Tong-Yee Lee
In this work, we introduce the notion of image retargetability to describe how well a particular image can be handled by content-aware image retargeting.