1 code implementation • 22 Nov 2023 • Yixin Liu, Chenrui Fan, Yutong Dai, Xun Chen, Pan Zhou, Lichao Sun
To solve these challenges, we propose MetaCloak, which solves the bi-level poisoning problem with a meta-learning framework with an additional transformation sampling process to craft transferable and robust perturbation.
no code implementations • 18 Sep 2023 • Taihui Li, Anish Lahiri, Yutong Dai, Owen Mayer
Demosaicing and denoising of RAW images are crucial steps in the processing pipeline of modern digital cameras.
1 code implementation • 8 Mar 2023 • Kai Zhang, Yutong Dai, Hongyi Wang, Eric Xing, Xun Chen, Lichao Sun
Federated learning is a promising paradigm that allows multiple clients to collaboratively train a model without sharing the local data.
1 code implementation • 7 Mar 2023 • Yihan Cao, Siyu Li, Yixin Liu, Zhiling Yan, Yutong Dai, Philip S. Yu, Lichao Sun
The goal of AIGC is to make the content creation process more efficient and accessible, allowing for the production of high-quality content at a faster pace.
1 code implementation • 27 Dec 2022 • Zixuan Ye, Yutong Dai, Chaoyi Hong, Zhiguo Cao, Hao Lu
Inspired by this, we introduce a novel composition style that binds the source and combined foregrounds in a definite triplet.
1 code implementation • 6 Dec 2022 • Yutong Dai, Zeyuan Chen, Junnan Li, Shelby Heinecke, Lichao Sun, ran Xu
We propose FedNH, a novel method that improves the local models' performance for both personalization and generalization by combining the uniformity and semantics of class prototypes.
2 code implementations • Genome Biology 2022 • Junru Jin, Yingying Yu, Ruheng Wang, Xin Zeng, Chao Pang, Yi Jiang, Zhongshen Li, Yutong Dai, Ran Su, Quan Zou, Kenta Nakai, Leyi Wei
In this study, we propose iDNA-ABF, a multi-scale deep biological language learning model that enables the interpretable prediction of DNA methylations based on genomic sequences only.
no code implementations • CVPR 2022 • Yutong Dai, Brian Price, He Zhang, Chunhua Shen
Deep image matting methods have achieved increasingly better results on benchmarks (e. g., Composition-1k/alphamatting. com).
no code implementations • 29 Nov 2021 • Dezhong Yao, Wanning Pan, Michael J O'Neill, Yutong Dai, Yao Wan, Hai Jin, Lichao Sun
To this end, this paper proposes FedHM, a novel heterogeneous federated model compression framework, distributing the heterogeneous low-rank models to clients and then aggregating them into a full-rank model.
no code implementations • 29 Sep 2021 • Yutong Dai, Xingjun Ma, Lichao Sun
Federated learning (FL) is a privacy-aware collaborative learning paradigm that allows multiple parties to jointly train a machine learning model without sharing their private data.
no code implementations • 30 Jun 2021 • Dezhong Yao, Wanning Pan, Yutong Dai, Yao Wan, Xiaofeng Ding, Hai Jin, Zheng Xu, Lichao Sun
Federated learning enables multiple clients to collaboratively learn a global model by periodically aggregating the clients' models without transferring the local data.
1 code implementation • CVPR 2021 • Yutong Dai, Hao Lu, Chunhua Shen
By looking at existing upsampling operators from a unified mathematical perspective, we generalize them into a second-order form and introduce Affinity-Aware Upsampling (A2U) where upsampling kernels are generated using a light-weight lowrank bilinear model and are conditioned on second-order features.
1 code implementation • 29 Jul 2020 • Frank E. Curtis, Yutong Dai, Daniel P. Robinson
We consider the problem of minimizing an objective function that is the sum of a convex function and a group sparsity-inducing regularizer.
Optimization and Control 49M37, 65K05, 65K10, 65Y20, 68Q25, 90C30, 90C60
no code implementations • 19 Aug 2019 • Yu Liu, Yutong Dai, Anh-Dzung Doan, Lingqiao Liu, Ian Reid
Through adding a common module, video loss, which we formulate with various forms of constraints (including weighted BCE loss, high-dimensional triplet loss, as well as a novel mixed instance-aware video loss), to train the parent network in the step (2), the network is then better prepared for the step (3), i. e. online fine-tuning on the target instance.
6 code implementations • 11 Aug 2019 • Hao Lu, Yutong Dai, Chunhua Shen, Songcen Xu
By viewing the indices as a function of the feature map, we introduce the concept of "learning to index", and present a novel index-guided encoder-decoder framework where indices are self-learned adaptively from data and are used to guide the downsampling and upsampling stages, without extra training supervision.
Ranked #2 on Grayscale Image Denoising on Set12 sigma30
1 code implementation • ICCV 2019 • Hao Lu, Yutong Dai, Chunhua Shen, Songcen Xu
We show that existing upsampling operators can be unified with the notion of the index function.