Search Results for author: Yijun Li

Found 21 papers, 11 papers with code

Computational Methods for Single-Cell Multi-Omics Integration and Alignment

no code implementations18 Jan 2022 Stefan Stanojevic, Yijun Li, Lana X. Garmire

Recently developed technologies to generate single-cell genomic data have made a revolutionary impact in the field of biology.

Machine Translation Translation

IMAGINE: Image Synthesis by Image-Guided Model Inversion

no code implementations CVPR 2021 Pei Wang, Yijun Li, Krishna Kumar Singh, Jingwan Lu, Nuno Vasconcelos

We introduce an inversion based method, denoted as IMAge-Guided model INvErsion (IMAGINE), to generate high-quality and diverse images from only a single training sample.

Image Generation

Few-shot Image Generation via Cross-domain Correspondence

1 code implementation CVPR 2021 Utkarsh Ojha, Yijun Li, Jingwan Lu, Alexei A. Efros, Yong Jae Lee, Eli Shechtman, Richard Zhang

Training generative models, such as GANs, on a target domain containing limited examples (e. g., 10) can easily result in overfitting.

Image Generation

Rethinking and Improving the Robustness of Image Style Transfer

1 code implementation CVPR 2021 Pei Wang, Yijun Li, Nuno Vasconcelos

Extensive research in neural style transfer methods has shown that the correlation between features extracted by a pre-trained VGG network has a remarkable ability to capture the visual style of an image.

Style Transfer

Content-Aware GAN Compression

1 code implementation CVPR 2021 Yuchen Liu, Zhixin Shu, Yijun Li, Zhe Lin, Federico Perazzi, S. Y. Kung

We then propose a novel content-aware method to guide the processes of both pruning and distillation.

Image Generation Image Manipulation +1

Neural Partial Differential Equations with Functional Convolution

no code implementations1 Jan 2021 Ziqian Wu, Xingzhe He, Michael Zhang, Yijun Li, Cheng Yang, Rui Liu, Shiying Xiong, Bo Zhu

Identifying the underlying structures of a PDE system based upon a small set of data samples on the solution space is challenging for machine learning.

Few-shot Image Generation with Elastic Weight Consolidation

no code implementations NeurIPS 2020 Yijun Li, Richard Zhang, Jingwan Lu, Eli Shechtman

Few-shot image generation seeks to generate more data of a given domain, with only few available training examples.

Image Generation

Learning to Caricature via Semantic Shape Transform

1 code implementation12 Aug 2020 Wenqing Chu, Wei-Chih Hung, Yi-Hsuan Tsai, Yu-Ting Chang, Yijun Li, Deng Cai, Ming-Hsuan Yang

Caricature is an artistic drawing created to abstract or exaggerate facial features of a person.


Modeling Artistic Workflows for Image Generation and Editing

1 code implementation ECCV 2020 Hung-Yu Tseng, Matthew Fisher, Jingwan Lu, Yijun Li, Vladimir Kim, Ming-Hsuan Yang

People often create art by following an artistic workflow involving multiple stages that inform the overall design.

Image Generation

Collaborative Distillation for Ultra-Resolution Universal Style Transfer

1 code implementation CVPR 2020 Huan Wang, Yijun Li, Yuehai Wang, Haoji Hu, Ming-Hsuan Yang

In this work, we present a new knowledge distillation method (named Collaborative Distillation) for encoder-decoder based neural style transfer to reduce the convolutional filters.

Knowledge Distillation Style Transfer

Controllable and Progressive Image Extrapolation

no code implementations25 Dec 2019 Yijun Li, Lu Jiang, Ming-Hsuan Yang

Image extrapolation aims at expanding the narrow field of view of a given image patch.

Im2Pencil: Controllable Pencil Illustration from Photographs

1 code implementation CVPR 2019 Yijun Li, Chen Fang, Aaron Hertzmann, Eli Shechtman, Ming-Hsuan Yang

We propose a high-quality photo-to-pencil translation method with fine-grained control over the drawing style.


Flow-Grounded Spatial-Temporal Video Prediction from Still Images

1 code implementation ECCV 2018 Yijun Li, Chen Fang, Jimei Yang, Zhaowen Wang, Xin Lu, Ming-Hsuan Yang

Existing video prediction methods mainly rely on observing multiple historical frames or focus on predicting the next one-frame.

Frame Video Prediction

A Closed-form Solution to Photorealistic Image Stylization

12 code implementations ECCV 2018 Yijun Li, Ming-Yu Liu, Xueting Li, Ming-Hsuan Yang, Jan Kautz

Photorealistic image stylization concerns transferring style of a reference photo to a content photo with the constraint that the stylized photo should remain photorealistic.

Image Stylization

Joint Image Filtering with Deep Convolutional Networks

no code implementations11 Oct 2017 Yijun Li, Jia-Bin Huang, Narendra Ahuja, Ming-Hsuan Yang

In contrast to existing methods that consider only the guidance image, the proposed algorithm can selectively transfer salient structures that are consistent with both guidance and target images.

Universal Style Transfer via Feature Transforms

15 code implementations NeurIPS 2017 Yijun Li, Chen Fang, Jimei Yang, Zhaowen Wang, Xin Lu, Ming-Hsuan Yang

The whitening and coloring transforms reflect a direct matching of feature covariance of the content image to a given style image, which shares similar spirits with the optimization of Gram matrix based cost in neural style transfer.

Image Reconstruction Style Transfer

Generative Face Completion

1 code implementation CVPR 2017 Yijun Li, Sifei Liu, Jimei Yang, Ming-Hsuan Yang

In this paper, we propose an effective face completion algorithm using a deep generative model.

Facial Inpainting Semantic Parsing

Diversified Texture Synthesis with Feed-forward Networks

no code implementations CVPR 2017 Yijun Li, Chen Fang, Jimei Yang, Zhaowen Wang, Xin Lu, Ming-Hsuan Yang

Recent progresses on deep discriminative and generative modeling have shown promising results on texture synthesis.

Texture Synthesis

Robust High Quality Image Guided Depth Upsampling

no code implementations17 Jun 2015 Wei Liu, Yijun Li, Xiaogang Chen, Jie Yang, Qiang Wu, Jingyi Yu

A popular solution is upsampling the obtained noisy low resolution depth map with the guidance of the companion high resolution color image.


Cannot find the paper you are looking for? You can Submit a new open access paper.