2 code implementations • 22 Nov 2024 • Zhenxiong Tan, Songhua Liu, Xingyi Yang, Qiaochu Xue, Xinchao Wang
In this paper, we introduce OminiControl, a highly versatile and parameter-efficient framework that integrates image conditions into pre-trained Diffusion Transformer (DiT) models.
no code implementations • 10 Oct 2024 • Ruonan Yu, Songhua Liu, Jingwen Ye, Xinchao Wang
Addressing these concerns, this paper introduces Teddy, a Taylor-approximated dataset distillation framework designed to handle large-scale dataset and enhance efficiency.
1 code implementation • 3 Sep 2024 • Songhua Liu, Weihao Yu, Zhenxiong Tan, Xinchao Wang
Modern diffusion models, particularly those utilizing a Transformer-based UNet for denoising, rely heavily on self-attention operations to manage complex spatial relationships, thus achieving impressive generation performance.
no code implementations • 15 Aug 2024 • Ruonan Yu, Songhua Liu, Zigeng Chen, Jingwen Ye, Xinchao Wang
Extensive experiments demonstrate that with only about 0. 003% of the original storage required for a complete set of soft labels, we achieve comparable performance to current state-of-the-art dataset distillation methods on large-scale datasets.
no code implementations • 24 Jun 2024 • Zhenxiong Tan, Xingyi Yang, Songhua Liu, Xinchao Wang
Specifically, we propose two coherent mechanisms: Clip parallelism and Dual-scope attention.
no code implementations • CVPR 2024 • Jingwen Ye, Ruonan Yu, Songhua Liu, Xinchao Wang
To investigate the impact of changes in training data on a pre-trained model, a common approach is leave-one-out retraining.
1 code implementation • CVPR 2024 • Shizun Wang, Songhua Liu, Zhenxiong Tan, Xinchao Wang
Currently, brain decoding is confined to a per-subject-per-model paradigm, limiting its applicability to the same individual for whom the decoding model is trained.
no code implementations • 20 Dec 2023 • Jingwen Ye, Ruonan Yu, Songhua Liu, Xinchao Wang
Our approach outperforms state-of-the-art attack methods and can be readily deployed as a plug-and-play solution.
1 code implementation • ICCV 2023 • Sucheng Ren, Xingyi Yang, Songhua Liu, Xinchao Wang
At the heart of our approach is to utilize a significance map, which is estimated through hybrid-scale self-attention and evolves itself during training, to reallocate tokens based on the significance of each region.
1 code implementation • CVPR 2023 • Runpeng Yu, Songhua Liu, Xingyi Yang, Xinchao Wang
Machine learning society has witnessed the emergence of a myriad of Out-of-Distribution (OoD) algorithms, which address the distribution shift between the training and the testing distribution by searching for a unified predictor or invariant feature representation.
1 code implementation • CVPRW 2023 • Marcos V. Conde, Manuel Kolmet, Tim Seizinger, Tom E. Bishop, Radu Timofte, Xiangyu Kong, Dafeng Zhang, Jinlong Wu, Fan Wang, Juewen Peng, Zhiyu Pan, Chengxin Liu, Xianrui Luo, Huiqiang Sun, Liao Shen, Zhiguo Cao, Ke Xian, Chaowei Liu, Zigeng Chen, Xingyi Yang, Songhua Liu, Yongcheng Jing, Michael Bi Mi, Xinchao Wang, Zhihao Yang, Wenyi Lian, Siyuan Lai, Haichuan Zhang, Trung Hoang, Amirsaeed Yazdani, Vishal Monga, Ziwei Luo, Fredrik K. Gustafsson, Zheng Zhao, Jens Sjölund, Thomas B. Schön, Yuxuan Zhao, Baoliang Chen, Yiqing Xu, JiXiang Niu
We present the new Bokeh Effect Transformation Dataset (BETD), and review the proposed solutions for this novel task at the NTIRE 2023 Bokeh Effect Transformation Challenge.
no code implementations • CVPR 2023 • Hao Tang, Songhua Liu, Tianwei Lin, Shaoli Huang, Fu Li, Dongliang He, Xinchao Wang
On the other hand, different from the vanilla version, we adopt a learnable scaling operation on content features before content-style feature interaction, which better preserves the original similarity between a pair of content features while ensuring the stylization quality.
1 code implementation • 19 Apr 2023 • Songhua Liu, Jingwen Ye, Xinchao Wang
Existing approaches either apply the holistic style of the style image in a global manner, or migrate local colors and textures of the style image to the content counterparts in a pre-defined way.
1 code implementation • CVPR 2023 • Jingwen Ye, Songhua Liu, Xinchao Wang
Unlike prior methods that update all or at least part of the parameters in the target network throughout the knowledge transfer process, PNC conducts partial parametric "cloning" from a source network and then injects the cloned module to the target, without modifying its parameters.
1 code implementation • 17 Jan 2023 • Ruonan Yu, Songhua Liu, Xinchao Wang
Recent success of deep learning is largely attributed to the sheer amount of data used for training deep neural networks. Despite the unprecedented success, the massive data, unfortunately, significantly increases the burden on storage and transmission and further gives rise to a cumbersome model training process.
1 code implementation • ICCV 2023 • Songhua Liu, Xinchao Wang
We pre-train the translator on some large datasets like ImageNet so that it requires only a limited number of adaptation steps on the target dataset.
no code implementations • CVPR 2023 • Songhua Liu, Jingwen Ye, Runpeng Yu, Xinchao Wang
In this paper, we explore the problem of slimmable dataset condensation, to extract a smaller synthetic dataset given only previous condensation results.
1 code implementation • NIPS 2022 • Songhua Liu, Kai Wang, Xingyi Yang, Jingwen Ye, Xinchao Wang
In this paper, we study dataset distillation (DD), from a novel perspective and introduce a \emph{dataset factorization} approach, termed \emph{HaBa}, which is a plug-and-play strategy portable to any existing DD baseline.
3 code implementations • 30 Oct 2022 • Songhua Liu, Kai Wang, Xingyi Yang, Jingwen Ye, Xinchao Wang
In this paper, we study \xw{dataset distillation (DD)}, from a novel perspective and introduce a \emph{dataset factorization} approach, termed \emph{HaBa}, which is a plug-and-play strategy portable to any existing DD baseline.
1 code implementation • 24 Oct 2022 • Xingyi Yang, Daquan Zhou, Songhua Liu, Jingwen Ye, Xinchao Wang
Given a collection of heterogeneous models pre-trained from distinct sources and with diverse architectures, the goal of DeRy, as its name implies, is to first dissect each model into distinctive building blocks, and then selectively reassemble the derived blocks to produce customized networks under both the hardware resource and performance constraints.
1 code implementation • 17 Jul 2022 • Jingwen Ye, Yifang Fu, Jie Song, Xingyi Yang, Songhua Liu, Xin Jin, Mingli Song, Xinchao Wang
Life-long learning aims at learning a sequence of tasks without forgetting the previously acquired knowledge.
1 code implementation • 13 Jul 2022 • Songhua Liu, Jingwen Ye, Sucheng Ren, Xinchao Wang
Prior approaches, despite the promising results, have relied on either estimating dense attention to compute per-point matching, which is limited to only coarse scales due to the quadratic memory cost, or fixing the number of correspondences to achieve linear complexity, which lacks flexibility.
2 code implementations • ICCV 2021 • Songhua Liu, Tianwei Lin, Dongliang He, Fu Li, Ruifeng Deng, Xin Li, Errui Ding, Hao Wang
Neural painting refers to the procedure of producing a series of strokes for a given image and non-photo-realistically recreating it using neural networks.
Ranked #1 on Object Detection on A2D
3 code implementations • ICCV 2021 • Songhua Liu, Tianwei Lin, Dongliang He, Fu Li, Meiling Wang, Xin Li, Zhengxing Sun, Qian Li, Errui Ding
Finally, the content feature is normalized so that they demonstrate the same local feature statistics as the calculated per-point weighted style feature statistics.
Ranked #5 on Style Transfer on StyleBench