Search Results for author: WeiMing Dong

Found 31 papers, 21 papers with code

Z-STAR+: A Zero-shot Style Transfer Method via Adjusting Style Distribution

no code implementations28 Nov 2024 Yingying Deng, Xiangyu He, Fan Tang, WeiMing Dong

In contrast to existing approaches, we have discovered that latent features in vanilla diffusion models inherently contain natural style and content distributions.

Contrastive Learning Image Denoising +1

LumiSculpt: A Consistency Lighting Control Network for Video Generation

no code implementations30 Oct 2024 Yuxin Zhang, Dandan Zheng, Biao Gong, Jingdong Chen, Ming Yang, WeiMing Dong, Changsheng Xu

Lighting plays a pivotal role in ensuring the naturalness of video generation, significantly influencing the aesthetic quality of the generated content.

Video Generation

Semantic Editing Increment Benefits Zero-Shot Composed Image Retrieval

1 code implementation ACM MM 2024 Zhenyu Yang, Shengsheng Qian, Dizhan Xue, JiaHong Wu, Fan Yang, WeiMing Dong, Changsheng Xu

To address this limitation, this paper proposes a training-free method called Semantic Editing Increment for ZS-CIR (SEIZE) to retrieve the target image based on the query image and text without training.

Image Retrieval Image to text +2

A Comprehensive Review of Few-shot Action Recognition

no code implementations20 Jul 2024 Yuyang Wanyan, Xiaoshan Yang, WeiMing Dong, Changsheng Xu

Few-shot action recognition aims to address the high cost and impracticality of manually labeling complex and variable video data in action recognition.

Few-Shot action recognition Few Shot Action Recognition +5

LDRE: LLM-based Divergent Reasoning and Ensemble for Zero-Shot Composed Image Retrieval

1 code implementation SIGIR 2024 Zhenyu Yang, Dizhan Xue, Shengsheng Qian, WeiMing Dong, Changsheng Xu

To conduct ZS-CIR, the prevailing methods employ pre-trained image-to-text models to transform the query image and text into a single text, which is then projected into the common feature space by CLIP to retrieve the target image.

Image Retrieval Image to text +2

Revealing the Two Sides of Data Augmentation: An Asymmetric Distillation-based Win-Win Solution for Open-Set Recognition

no code implementations28 Apr 2024 Yunbing Jia, Xiaoyu Kong, Fan Tang, Yixing Gao, WeiMing Dong, Yi Yang

In this paper, we reveal the two sides of data augmentation: enhancements in closed-set recognition correlate with a significant decrease in open-set recognition.

Data Augmentation Knowledge Distillation +1

Break-for-Make: Modular Low-Rank Adaptations for Composable Content-Style Customization

no code implementations28 Mar 2024 Yu Xu, Fan Tang, Juan Cao, Yuxin Zhang, Oliver Deussen, WeiMing Dong, Jintao Li, Tong-Yee Lee

Based on the adapters broken apart for separate training content and style, we then make the entity parameter space by reconstructing the content and style PLPs matrices, followed by fine-tuning the combined adapter to generate the target object with the desired appearance.

CreativeSynth: Creative Blending and Synthesis of Visual Arts based on Multimodal Diffusion

1 code implementation25 Jan 2024 Nisha Huang, WeiMing Dong, Yuxin Zhang, Fan Tang, Ronghui Li, Chongyang Ma, Xiu Li, Changsheng Xu

Large-scale text-to-image generative models have made impressive strides, showcasing their ability to synthesize a vast array of high-quality images.

Image Generation Style Transfer

Z*: Zero-shot Style Transfer via Attention Reweighting

1 code implementation CVPR 2024 Yingying Deng, Xiangyu He, Fan Tang, WeiMing Dong

Despite the remarkable progress in image style transfer formulating style in the context of art is inherently subjective and challenging.

Image Denoising Style Transfer

MotionCrafter: One-Shot Motion Customization of Diffusion Models

1 code implementation8 Dec 2023 Yuxin Zhang, Fan Tang, Nisha Huang, Haibin Huang, Chongyang Ma, WeiMing Dong, Changsheng Xu

The essence of a video lies in its dynamic motions, including character actions, object movements, and camera movements.

Disentanglement Motion Disentanglement +3

$Z^*$: Zero-shot Style Transfer via Attention Rearrangement

1 code implementation25 Nov 2023 Yingying Deng, Xiangyu He, Fan Tang, WeiMing Dong

Despite the remarkable progress in image style transfer, formulating style in the context of art is inherently subjective and challenging.

Style Transfer

ProSpect: Prompt Spectrum for Attribute-Aware Personalization of Diffusion Models

3 code implementations25 May 2023 Yuxin Zhang, WeiMing Dong, Fan Tang, Nisha Huang, Haibin Huang, Chongyang Ma, Tong-Yee Lee, Oliver Deussen, Changsheng Xu

We apply ProSpect in various personalized attribute-aware image generation applications, such as image-guided or text-driven manipulations of materials, style, and layout, achieving previously unattainable results from a single image input without fine-tuning the diffusion models.

Attribute Disentanglement +1

Style-A-Video: Agile Diffusion for Arbitrary Text-based Video Style Transfer

1 code implementation9 May 2023 Nisha Huang, Yuxin Zhang, WeiMing Dong

Large-scale text-to-video diffusion models have demonstrated an exceptional ability to synthesize diverse videos.

Denoising Style Transfer +1

A Unified Arbitrary Style Transfer Framework via Adaptive Contrastive Learning

1 code implementation9 Mar 2023 Yuxin Zhang, Fan Tang, WeiMing Dong, Haibin Huang, Chongyang Ma, Tong-Yee Lee, Changsheng Xu

Our framework consists of three key components, i. e., a parallel contrastive learning scheme for style representation and style transfer, a domain enhancement module for effective learning of style distribution, and a generative network for style transfer.

Contrastive Learning Representation Learning +1

Region-Aware Diffusion for Zero-shot Text-driven Image Editing

1 code implementation23 Feb 2023 Nisha Huang, Fan Tang, WeiMing Dong, Tong-Yee Lee, Changsheng Xu

Different from current mask-based image editing methods, we propose a novel region-aware diffusion model (RDM) for entity-level image editing, which could automatically locate the region of interest and replace it following given text prompts.

Image Manipulation

Inversion-Based Style Transfer with Diffusion Models

1 code implementation CVPR 2023 Yuxin Zhang, Nisha Huang, Fan Tang, Haibin Huang, Chongyang Ma, WeiMing Dong, Changsheng Xu

Our key idea is to learn artistic style directly from a single painting and then guide the synthesis without providing complex textual descriptions.

Denoising Style Transfer +1

DiffStyler: Controllable Dual Diffusion for Text-Driven Image Stylization

1 code implementation19 Nov 2022 Nisha Huang, Yuxin Zhang, Fan Tang, Chongyang Ma, Haibin Huang, Yong Zhang, WeiMing Dong, Changsheng Xu

Despite the impressive results of arbitrary image-guided style transfer methods, text-driven image stylization has recently been proposed for transferring a natural image into a stylized one according to textual descriptions of the target style provided by the user.

Denoising Image Stylization

Understanding and Mitigating Overfitting in Prompt Tuning for Vision-Language Models

1 code implementation4 Nov 2022 Chengcheng Ma, Yang Liu, Jiankang Deng, Lingxi Xie, WeiMing Dong, Changsheng Xu

Pretrained vision-language models (VLMs) such as CLIP have shown impressive generalization capability in downstream vision tasks with appropriate text prompts.

object-detection Open Vocabulary Object Detection +2

Draw Your Art Dream: Diverse Digital Art Synthesis with Multimodal Guided Diffusion

1 code implementation27 Sep 2022 Nisha Huang, Fan Tang, WeiMing Dong, Changsheng Xu

Extensive experimental results on the quality and quantity of the generated digital art paintings confirm the effectiveness of the combination of the diffusion model and multimodal guidance.

Diversity

Domain Enhanced Arbitrary Image Style Transfer via Contrastive Learning

1 code implementation19 May 2022 Yuxin Zhang, Fan Tang, WeiMing Dong, Haibin Huang, Chongyang Ma, Tong-Yee Lee, Changsheng Xu

Our framework consists of three key components, i. e., a multi-layer style projector for style code encoding, a domain enhancement module for effective learning of style distribution, and a generative network for image style transfer.

Contrastive Learning Image Stylization +1

StyTr2: Image Style Transfer With Transformers

3 code implementations CVPR 2022 Yingying Deng, Fan Tang, WeiMing Dong, Chongyang Ma, Xingjia Pan, Lei Wang, Changsheng Xu

The goal of image style transfer is to render an image with artistic features guided by a style reference while maintaining the original content.

Decoder Style Transfer

Evo-ViT: Slow-Fast Token Evolution for Dynamic Vision Transformer

1 code implementation3 Aug 2021 Yifan Xu, Zhijie Zhang, Mengdan Zhang, Kekai Sheng, Ke Li, WeiMing Dong, Liqing Zhang, Changsheng Xu, Xing Sun

Vision transformers (ViTs) have recently received explosive popularity, but the huge computational cost is still a severe issue.

Efficient ViTs

User-Guided Personalized Image Aesthetic Assessment based on Deep Reinforcement Learning

no code implementations14 Jun 2021 Pei Lv, Jianqi Fan, Xixi Nie, WeiMing Dong, Xiaoheng Jiang, Bing Zhou, Mingliang Xu, Changsheng Xu

This framework leverages user interactions to retouch and rank images for aesthetic assessment based on deep reinforcement learning (DRL), and generates personalized aesthetic distribution that is more in line with the aesthetic preferences of different users.

Deep Reinforcement Learning Image Enhancement +2

StyTr$^2$: Image Style Transfer with Transformers

4 code implementations30 May 2021 Yingying Deng, Fan Tang, WeiMing Dong, Chongyang Ma, Xingjia Pan, Lei Wang, Changsheng Xu

The goal of image style transfer is to render an image with artistic features guided by a style reference while maintaining the original content.

Decoder Style Transfer

Towards Corruption-Agnostic Robust Domain Adaptation

no code implementations21 Apr 2021 Yifan Xu, Kekai Sheng, WeiMing Dong, Baoyuan Wu, Changsheng Xu, Bao-Gang Hu

However, due to unpredictable corruptions (e. g., noise and blur) in real data like web images, domain adaptation methods are increasingly required to be corruption robust on target domains.

Domain Adaptation

On Evolving Attention Towards Domain Adaptation

no code implementations25 Mar 2021 Kekai Sheng, Ke Li, Xiawu Zheng, Jian Liang, WeiMing Dong, Feiyue Huang, Rongrong Ji, Xing Sun

However, considering that the configuration of attention, i. e., the type and the position of attention module, affects the performance significantly, it is more generalized to optimize the attention configuration automatically to be specialized for arbitrary UDA scenario.

Partial Domain Adaptation Unsupervised Domain Adaptation

Unveiling the Potential of Structure Preserving for Weakly Supervised Object Localization

1 code implementation CVPR 2021 Xingjia Pan, Yingguo Gao, Zhiwen Lin, Fan Tang, WeiMing Dong, Haolei Yuan, Feiyue Huang, Changsheng Xu

Weakly supervised object localization(WSOL) remains an open problem given the deficiency of finding object extent information using a classification network.

Classification General Classification +3

Effective Label Propagation for Discriminative Semi-Supervised Domain Adaptation

no code implementations4 Dec 2020 Zhiyong Huang, Kekai Sheng, WeiMing Dong, Xing Mei, Chongyang Ma, Feiyue Huang, Dengwen Zhou, Changsheng Xu

For intra-domain propagation, we propose an effective self-training strategy to mitigate the noises in pseudo-labeled target domain data and improve the feature discriminability in the target domain.

Domain Adaptation Image Classification +1

Image Retargeting by Content-Aware Synthesis

no code implementations26 Mar 2014 Weiming Dong, Fuzhang Wu, Yan Kong, Xing Mei, Tong-Yee Lee, Xiaopeng Zhang

We propose to retarget the textural regions by content-aware synthesis and non-textural regions by fast multi-operators.

Image Retargeting

Cannot find the paper you are looking for? You can Submit a new open access paper.