no code implementations • 4 Feb 2025 • Senmao Li, Kai Wang, Joost Van de Weijer, Fahad Shahbaz Khan, Chun-Le Guo, Shiqi Yang, Yaxing Wang, Jian Yang, Ming-Ming Cheng
Diffusion priors have been used for blind face restoration (BFR) by fine-tuning diffusion models (DMs) on restoration datasets to recover low-quality images.
1 code implementation • 23 Jan 2025 • Tao Liu, Kai Wang, Senmao Li, Joost Van de Weijer, Fahad Shahbaz Khan, Shiqi Yang, Yaxing Wang, Jian Yang, Ming-Ming Cheng
Drawing inspiration from the inherent context consistency, we propose a novel training-free method for consistent text-to-image (T2I) generation, termed "One-Prompt-One-Story" (1Prompt1Story).
no code implementations • 21 Jan 2025 • Samantha Min Er Yew, Xiaofeng Lei, Jocelyn Hui Lin Goh, Yibing Chen, Sahana Srinivasan, Miao-li Chee, Krithi Pushpanathan, Ke Zou, Qingshan Hou, Zhi Da Soh, Cancan Xue, Marco Chak Yan Yu, Charumathi Sabanayagam, E Shyong Tai, Xueling Sim, Yaxing Wang, Jost B. Jonas, Vinay Nangia, Gabriel Dawei Yang, Emma Anran Ran, Carol Yim-Lui Cheung, Yangqin Feng, Jun Zhou, Rick Siow Mong Goh, Yukun Zhou, Pearse A. Keane, Yong liu, Ching-Yu Cheng, Yih-Chung Tham
Background: RETFound, a self-supervised, retina-specific foundation model (FM), showed potential in downstream applications.
no code implementations • 10 Jan 2025 • Minxing Luo, Zixun Xia, Liaojun Chen, Zhenhang Li, Weichao Zeng, Jianye Wang, Wentao Cheng, Yaxing Wang, Yu Zhou, Jian Yang
In this paper, we introduce a new training-free framework, STGen, which accurately generates visual texts in challenging scenarios (\eg, slanted or curved text layouts) while harmonizing them with the text background.
1 code implementation • 11 Nov 2024 • Taihang Hu, Linxuan Li, Joost Van de Weijer, Hongcheng Gao, Fahad Shahbaz Khan, Jian Yang, Ming-Ming Cheng, Kai Wang, Yaxing Wang
In this paper, we define semantic binding as the task of associating a given object with its attribute, termed attribute binding, or linking it to other related sub-objects, referred to as object binding.
no code implementations • 7 Nov 2024 • Yichen Shi, Zhuofu Tao, Yuhao Gao, Tianjia Zhou, Cheng Chang, Yaxing Wang, BingYu Chen, Genhao Zhang, Alvin Liu, Zhiping Yu, Ting-Jung Lin, Lei He
A significant portion of the effort is experience-driven, which makes the automation of AMS circuit design a formidable challenge.
1 code implementation • 8 Feb 2024 • Senmao Li, Joost Van de Weijer, Taihang Hu, Fahad Shahbaz Khan, Qibin Hou, Yaxing Wang, Jian Yang
However, these models struggle to effectively suppress the generation of undesired content, which is explicitly requested to be omitted from the generated image in the prompt.
1 code implementation • 15 Dec 2023 • Senmao Li, Taihang Hu, Joost Van de Weijer, Fahad Shahbaz Khan, Tao Liu, Linxuan Li, Shiqi Yang, Yaxing Wang, Ming-Ming Cheng, Jian Yang
This insight motivates us to omit encoder computation at certain adjacent time-steps and reuse encoder features of previous time-steps as input to the decoder in multiple time-steps.
1 code implementation • 12 Dec 2023 • Kangneng Zhou, Daiheng Gao, Xuan Wang, Jie Zhang, Peng Zhang, Xusen Sun, Longhao Zhang, Shiqi Yang, Bang Zhang, Liefeng Bo, Yaxing Wang, Ming-Ming Cheng
This enhances masked-based editing in local areas; second, we present a novel distillation strategy: Conditional Distillation on Geometry and Texture (CDGT).
no code implementations • 8 Sep 2023 • Yupeng Zhou, Daquan Zhou, Zuo-Liang Zhu, Yaxing Wang, Qibin Hou, Jiashi Feng
In this work, we identify that a crucial factor leading to the text-image mismatch issue is the inadequate cross-modality relation learning between the prompt and the output image.
no code implementations • 1 Sep 2023 • Shiqi Yang, Yaxing Wang, Joost Van de Weijer, Luis Herranz, Shangling Jui, Jian Yang
We capture this intrinsic structure by defining local affinity of the target data, and encourage label consistency among data with high local affinity.
1 code implementation • 14 May 2023 • Dixian Zhu, Bokun Wang, Zhi Chen, Yaxing Wang, Milan Sonka, Xiaodong Wu, Tianbao Yang
This paper considers a novel application of deep AUC maximization (DAM) for multi-instance learning (MIL), in which a single class label is assigned to a bag of instances (e. g., multiple 2D slices of a CT scan for a patient).
1 code implementation • 28 Mar 2023 • Senmao Li, Joost Van de Weijer, Taihang Hu, Fahad Shahbaz Khan, Qibin Hou, Yaxing Wang, Jian Yang, Ming-Ming Cheng
A significant research effort is focused on exploiting the amazing capacities of pretrained diffusion models for the editing of images. They either finetune the model, or invert the image in the latent space of the pretrained model.
Ranked #9 on
Text-based Image Editing
on PIE-Bench
1 code implementation • CVPR 2023 • Senmao Li, Joost Van de Weijer, Yaxing Wang, Fahad Shahbaz Khan, Meiqin Liu, Jian Yang
In the second step, based on the well-trained multi-class 3D-aware GAN architecture, that preserves view-consistency, we construct a 3D-aware I2I translation system.
1 code implementation • 6 Mar 2023 • Xinhui Li, Mingjia Li, Yaxing Wang, Chuan-Xian Ren, Xiaojie Guo
Domain generalization in semantic segmentation aims to alleviate the performance degradation on unseen domains through learning domain-invariant features.
1 code implementation • 7 Jun 2022 • Shiqi Yang, Yaxing Wang, Kai Wang, Shangling Jui, Joost Van de Weijer
In this paper, we investigate Source-free Open-partial Domain Adaptation (SF-OPDA), which addresses the situation where there exist both domain and category shifts between source and target domains.
no code implementations • 30 May 2022 • Aitor Alvarez-Gila, Joost Van de Weijer, Yaxing Wang, Estibaliz Garrote
We present MVMO (Multi-View, Multi-Object dataset): a synthetic dataset of 116, 000 scenes containing randomly placed objects of 10 distinct classes and captured from 25 camera locations in the upper hemisphere.
1 code implementation • 9 May 2022 • Shiqi Yang, Yaxing Wang, Kai Wang, Shangling Jui, Joost Van de Weijer
Treating SFDA as an unsupervised clustering problem and following the intuition that local neighbors in feature space should have more similar predictions than other features, we propose to optimize an objective of prediction consistency.
1 code implementation • 4 Dec 2021 • Héctor Laria, Yaxing Wang, Joost Van de Weijer, Bogdan Raducanu
GANs have matured in recent years and are able to generate high-resolution, realistic images.
no code implementations • 25 Nov 2021 • Fei Yang, Yaxing Wang, Luis Herranz, Yongmei Cheng, Mikhail Mozerov
Thus, we further propose a unified framework that allows both translation and autoencoding capabilities in a single codec.
2 code implementations • NeurIPS 2021 • Shiqi Yang, Yaxing Wang, Joost Van de Weijer, Luis Herranz, Shangling Jui
In this paper, we address the challenging source-free domain adaptation (SFDA) problem, where the source pretrained model is adapted to the target domain in the absence of source data.
Ranked #7 on
Source-Free Domain Adaptation
on VisDA-2017
no code implementations • ICLR 2022 • Yaxing Wang, Joost Van de Weijer, Lu Yu, Shangling Jui
Therefore, we investigate knowledge distillation to transfer knowledge from a high-quality unconditioned generative model (e. g., StyleGAN) to a conditioned synthetic image generation modules in a variety of systems.
1 code implementation • ICCV 2021 • Shiqi Yang, Yaxing Wang, Joost Van de Weijer, Luis Herranz, Shangling Jui
In this paper, we propose a new domain adaptation paradigm called Generalized Source-free Domain Adaptation (G-SFDA), where the learned model needs to perform well on both the target and source domains, with only access to current unlabeled target data during adaptation.
Ranked #8 on
Source-Free Domain Adaptation
on VisDA-2017
no code implementations • ICCV 2021 • Yaxing Wang, Hector Laria Mantecon, Joost Van de Weijer, Laura Lopez-Fuentes, Bogdan Raducanu
In this paper, we propose a new transfer learning for I2I translation (TransferI2I).
1 code implementation • 28 Apr 2021 • Yaxing Wang, Abel Gonzalez-Garcia, Chenshen Wu, Luis Herranz, Fahad Shahbaz Khan, Shangling Jui, Joost Van de Weijer
Therefore, we propose a novel knowledge transfer method for generative models based on mining the knowledge that is most beneficial to a specific target domain, either from a single or multiple pretrained GANs.
1 code implementation • NeurIPS 2020 • Yaxing Wang, Lu Yu, Joost Van de Weijer
To enable the training of deep I2I models on small datasets, we propose a novel transfer learning method, that transfers knowledge from pre-trained GANs.
2 code implementations • 23 Oct 2020 • Shiqi Yang, Yaxing Wang, Joost Van de Weijer, Luis Herranz, Shangling Jui
When adapting to the target domain, the additional classifier initialized from source classifier is expected to find misclassified features.
Source-Free Domain Adaptation
Unsupervised Domain Adaptation
no code implementations • 2 Jul 2020 • Hui Xie, Zhe Pan, Leixin Zhou, Fahim A Zaman, Danny Chen, Jost B Jonas, Yaxing Wang, Xiaodong Wu
In this work, we propose to parameterize the surface cost functions in the graph model and leverage DL to learn those parameters.
1 code implementation • CVPR 2020 • Yaxing Wang, Salman Khan, Abel Gonzalez-Garcia, Joost Van de Weijer, Fahad Shahbaz Khan
In this work, we go one step further and reduce the amount of required labeled data also from the source domain during training.
3 code implementations • ECCV 2020 • Lei Kang, Pau Riba, Yaxing Wang, Marçal Rusiñol, Alicia Fornés, Mauricio Villegas
We propose a novel method that is able to produce credible handwritten word images by conditioning the generative process with both calligraphic style features and textual content.
2 code implementations • CVPR 2020 • Yaxing Wang, Abel Gonzalez-Garcia, David Berga, Luis Herranz, Fahad Shahbaz Khan, Joost Van de Weijer
We propose a novel knowledge transfer method for generative models based on mining the knowledge that is most beneficial to a specific target domain, either from a single or multiple pretrained GANs.
2 code implementations • 19 Aug 2019 • Yaxing Wang, Abel Gonzalez-Garcia, Joost Van de Weijer, Luis Herranz
Recently, image-to-image translation research has witnessed remarkable progress.
no code implementations • 23 Jul 2019 • Yaxing Wang, Abel Gonzalez-Garcia, Joost Van de Weijer, Luis Herranz
The task of unpaired image-to-image translation is highly challenging due to the lack of explicit cross-domain pairs of instances.
no code implementations • 8 Mar 2019 • Yaxing Wang, Luis Herranz, Joost Van de Weijer
This paper addresses the problem of inferring unseen cross-modal image-to-image translations between multiple modalities.
1 code implementation • NeurIPS 2018 • Chenshen Wu, Luis Herranz, Xialei Liu, Yaxing Wang, Joost Van de Weijer, Bogdan Raducanu
In particular, we investigate generative adversarial networks (GANs) in the task of learning new categories in a sequential fashion.
2 code implementations • 6 Sep 2018 • Chenshen Wu, Luis Herranz, Xialei Liu, Yaxing Wang, Joost Van de Weijer, Bogdan Raducanu
In particular, we investigate generative adversarial networks (GANs) in the task of learning new categories in a sequential fashion.
1 code implementation • ECCV 2018 • Yaxing Wang, Chenshen Wu, Luis Herranz, Joost Van de Weijer, Abel Gonzalez-Garcia, Bogdan Raducanu
Transferring the knowledge of pretrained networks to new domains by means of finetuning is a widely used practice for applications based on discriminative models.
Ranked #7 on
10-shot image generation
on Babies
1 code implementation • CVPR 2018 • Yaxing Wang, Joost Van de Weijer, Luis Herranz
We address the problem of image translation between domains or modalities for which no direct paired data is available (i. e. zero-pair translation).
no code implementations • 3 Dec 2016 • Yaxing Wang, Lichao Zhang, Joost Van de Weijer
The first one is based on the fact that in the minimax game which is played to optimize the GAN objective the generator network keeps on changing even after the network can be considered optimal.
1 code implementation • WS 2016 • Ozan Caglayan, Walid Aransa, Yaxing Wang, Marc Masana, Mercedes García-Martínez, Fethi Bougares, Loïc Barrault, Joost Van de Weijer
This paper presents the systems developed by LIUM and CVC for the WMT16 Multimodal Machine Translation challenge.