no code implementations • 6 Mar 2024 • Gen Li, Yu Huang, Timofey Efimov, Yuting Wei, Yuejie Chi, Yuxin Chen
Score-based diffusion models, while achieving remarkable empirical performance, often suffer from low sampling speed, due to extensive function evaluations needed during the sampling phase.
no code implementations • 12 Feb 2024 • Gen Li, Zhihan Huang, Yuting Wei
Consistency models, which were proposed to mitigate the high computational overhead during the sampling phase of diffusion models, facilitate single-step sampling while attaining state-of-the-art empirical performance.
no code implementations • 16 Jan 2024 • Gen Li, Kaifeng Zhao, Siwei Zhang, Xiaozhong Lyu, Mihai Dusmanu, Yan Zhang, Marc Pollefeys, Siyu Tang
To address this challenge, we introduce EgoGen, a new synthetic data generator that can produce accurate and rich ground-truth training data for egocentric perception tasks.
no code implementations • 8 Jan 2024 • Gen Li, Yuting Wei
Characterizing the distribution of high-dimensional statistical estimators is a challenging task, due to the breakdown of classical asymptotic theory in high dimension.
no code implementations • 22 Dec 2023 • Yicheng Leng, Chaowei Fang, Gen Li, Yixiang Fang, Guanbin Li
Visible watermarks, while instrumental in protecting image copyrights, frequently distort the underlying content, complicating tasks like scene interpretation and image editing.
no code implementations • 15 Dec 2023 • Yucong Dai, Gen Li, Feng Luo, Xiaolong Ma, Yongkai Wu
To address this, we define a fair pruning task where a sparse model is derived subject to fairness requirements.
no code implementations • 29 Nov 2023 • Gen Li, Deqing Sun, Laura Sevilla-Lara, Varun Jampani
We introduce One-shot Open Affordance Learning (OOAL), where a model is trained with just one example per base object category, but is expected to identify novel objects and affordances.
no code implementations • 10 Oct 2023 • Shreyank N Gowda, Xinyue Hao, Gen Li, Laura Sevilla-Lara, Shashank Narayana Gowda
Deep learning models have revolutionized various fields, from image recognition to natural language processing, by achieving unprecedented levels of accuracy.
no code implementations • 5 Sep 2023 • Shaohua Liu, Yu Qi, Gen Li, Mingjian Chen, Teng Zhang, Jia Cheng, Jun Lei
Specifically, we construct subgraphs of spatial, temporal, spatial-temporal, and global views respectively to precisely characterize the user's interests in various contexts.
1 code implementation • 5 Jul 2023 • Liangbin Xie, Xintao Wang, Xiangyu Chen, Gen Li, Ying Shan, Jiantao Zhou, Chao Dong
After detecting the artifact regions, we develop a finetune procedure to improve GAN-based SR models with a few samples, so that they can deal with similar types of artifacts in more unseen real data.
1 code implementation • 18 Jun 2023 • Jonghyun Kim, Gen Li, Joongkyu Kim
Despite recent progress in semantic image synthesis, complete control over image style remains a challenging problem.
no code implementations • 15 Jun 2023 • Gen Li, Yuting Wei, Yuxin Chen, Yuejie Chi
Diffusion models, which convert noise into new data instances by learning to reverse a Markov diffusion process, have become a cornerstone in contemporary generative modeling.
no code implementations • 30 May 2023 • Gen Li, Weichen Wu, Yuejie Chi, Cong Ma, Alessandro Rinaldo, Yuting Wei
This paper is concerned with the problem of policy evaluation with linear function approximation in discounted infinite horizon Markov decision processes.
no code implementations • NeurIPS 2023 • Laixi Shi, Gen Li, Yuting Wei, Yuxin Chen, Matthieu Geist, Yuejie Chi
Assuming access to a generative model that draws samples based on the nominal MDP, we characterize the sample complexity of RMDPs when the uncertainty set is specified via either the total variation (TV) distance or $\chi^2$ divergence.
no code implementations • NeurIPS 2023 • Xiang Ji, Gen Li
A crucial problem in reinforcement learning is learning the optimal policy.
no code implementations • 7 May 2023 • Gen Li, Ganghua Wang, Jie Ding
In this paper, the territory of LASSO is extended to two-layer ReLU neural networks, a fashionable and powerful nonlinear regression model.
no code implementations • 14 Apr 2023 • Gen Li, Yuling Yan, Yuxin Chen, Jianqing Fan
This paper studies reward-agnostic exploration in reinforcement learning (RL) -- a scenario where the learner is unware of the reward functions during the exploration stage -- and designs an algorithm that improves over the state of the art.
no code implementations • CVPR 2023 • Gen Li, Varun Jampani, Deqing Sun, Laura Sevilla-Lara
A key step to acquire this skill is to identify what part of the object affords each action, which is called affordance grounding.
1 code implementation • CVPR 2023 • Gen Li, Jie Ji, Minghai Qin, Wei Niu, Bin Ren, Fatemeh Afghah, Linke Guo, Xiaolong Ma
To reconcile such, we propose a novel method for high-quality and efficient video resolution upscaling tasks, which leverages the spatial-temporal information to accurately divide video into chunks, thus keeping the number of chunks as well as the model size to minimum.
no code implementations • 7 Feb 2023 • Gen Li, Wei Fan, Yuting Wei
This paper is concerned with the problem of reconstructing an unknown rank-one matrix with prior structural information from noisy observations.
1 code implementation • CVPR 2023 • Fanghua Yu, Xintao Wang, Mingdeng Cao, Gen Li, Ying Shan, Chao Dong
Omnidirectional images (ODIs) have obtained lots of research interest for immersive experiences.
no code implementations • 30 Jan 2023 • Gen Li, Yanxi Chen, Yuejie Chi, H. Vincent Poor, Yuxin Chen
Efficient computation of the optimal transport distance between two distributions serves as an algorithm subroutine that empowers various applications.
2 code implementations • 7 Nov 2022 • Andrey Ignatov, Radu Timofte, Maurizio Denna, Abdel Younes, Ganzorig Gankhuyag, Jingang Huh, Myeong Kyun Kim, Kihwan Yoon, Hyeon-Cheol Moon, Seungho Lee, Yoonsik Choe, Jinwoo Jeong, Sungjei Kim, Maciej Smyl, Tomasz Latkowski, Pawel Kubik, Michal Sokolski, Yujie Ma, Jiahao Chao, Zhou Zhou, Hongfan Gao, Zhengfeng Yang, Zhenbing Zeng, Zhengyang Zhuge, Chenghua Li, Dan Zhu, Mengdi Sun, Ran Duan, Yan Gao, Lingshun Kong, Long Sun, Xiang Li, Xingdong Zhang, Jiawei Zhang, Yaqi Wu, Jinshan Pan, Gaocheng Yu, Jin Zhang, Feng Zhang, Zhe Ma, Hongbin Wang, Hojin Cho, Steve Kim, Huaen Li, Yanbo Ma, Ziwei Luo, Youwei Li, Lei Yu, Zhihong Wen, Qi Wu, Haoqiang Fan, Shuaicheng Liu, Lize Zhang, Zhikai Zong, Jeremy Kwon, Junxi Zhang, Mengyuan Li, Nianxiang Fu, Guanchen Ding, Han Zhu, Zhenzhong Chen, Gen Li, Yuanfan Zhang, Lei Sun, Dafeng Zhang, Neo Yang, Fitz Liu, Jerry Zhao, Mustafa Ayazoglu, Bahri Batuhan Bilecen, Shota Hirose, Kasidis Arunruangsirilert, Luo Ao, Ho Chun Leung, Andrew Wei, Jie Liu, Qiang Liu, Dahai Yu, Ao Li, Lei Luo, Ce Zhu, Seongmin Hong, Dongwon Park, Joonhee Lee, Byeong Hyun Lee, Seunggyu Lee, Se Young Chun, Ruiyuan He, Xuhao Jiang, Haihang Ruan, Xinjian Zhang, Jing Liu, Garas Gendy, Nabil Sabor, Jingchao Hou, Guanghui He
While numerous solutions have been proposed for this problem in the past, they are usually not compatible with low-power mobile NPUs having many computational and memory constraints.
no code implementations • 6 Sep 2022 • Xi Wang, Gen Li, Yen-Ling Kuo, Muhammed Kocabas, Emre Aksan, Otmar Hilliges
We further qualitatively evaluate the effectiveness of our method on real images and demonstrate its generalizability towards interaction types and object categories.
no code implementations • 22 Aug 2022 • Gen Li, Yuejie Chi, Yuting Wei, Yuxin Chen
This paper studies multi-agent reinforcement learning in Markov games, with the goal of learning Nash equilibria or coarse correlated equilibria (CCE) sample-optimally.
no code implementations • 5 Aug 2022 • Gen Li, Yuting Wei
As two concrete consequences of the proposed analysis recipe: (i) when solving $\mathbb{Z}_2$ synchronization, we predict the behavior of spectrally initialized AMP for up to $O\big(\frac{n}{\mathrm{poly}\log n}\big)$ iterations, showing that the algorithm succeeds without the need of a subsequent refinement stage (as conjectured recently by \citet{celentano2021local}); (ii) we characterize the non-asymptotic behavior of AMP in sparse PCA (in the spiked Wigner model) for a broad range of signal-to-noise ratio.
no code implementations • 28 Jul 2022 • Yuanfan Zhang, Gen Li, Lei Sun
Since convolutional neural networks perform well in learning generalizable image priors from large-scale data, these models have been widely used in image denoising tasks.
no code implementations • 20 Jul 2022 • Aijin Li, Gen Li, Lei Sun, Xintao Wang
Blind face restoration usually encounters with diverse scale face inputs, especially in the real world.
1 code implementation • 14 Jun 2022 • Yanze Wu, Xintao Wang, Gen Li, Ying Shan
This paper studies the problem of real-world video super-resolution (VSR) for animation videos, and reveals three key improvements for practical animation VSR.
no code implementations • 8 Jun 2022 • Yuling Yan, Gen Li, Yuxin Chen, Jianqing Fan
This paper makes progress towards learning Nash equilibria in two-player zero-sum Markov games from offline data.
Model-based Reinforcement Learning reinforcement-learning +1
1 code implementation • 13 May 2022 • YuChao Gu, Xintao Wang, Liangbin Xie, Chao Dong, Gen Li, Ying Shan, Ming-Ming Cheng
Equipped with the VQ codebook as a facial detail dictionary and the parallel decoder design, the proposed VQFR can largely enhance the restored quality of facial details while keeping the fidelity to previous methods.
2 code implementations • 11 May 2022 • Yawei Li, Kai Zhang, Radu Timofte, Luc van Gool, Fangyuan Kong, Mingxi Li, Songwei Liu, Zongcai Du, Ding Liu, Chenhui Zhou, Jingyi Chen, Qingrui Han, Zheyuan Li, Yingqi Liu, Xiangyu Chen, Haoming Cai, Yu Qiao, Chao Dong, Long Sun, Jinshan Pan, Yi Zhu, Zhikai Zong, Xiaoxiao Liu, Zheng Hui, Tao Yang, Peiran Ren, Xuansong Xie, Xian-Sheng Hua, Yanbo Wang, Xiaozhong Ji, Chuming Lin, Donghao Luo, Ying Tai, Chengjie Wang, Zhizhong Zhang, Yuan Xie, Shen Cheng, Ziwei Luo, Lei Yu, Zhihong Wen, Qi Wu1, Youwei Li, Haoqiang Fan, Jian Sun, Shuaicheng Liu, Yuanfei Huang, Meiguang Jin, Hua Huang, Jing Liu, Xinjian Zhang, Yan Wang, Lingshun Long, Gen Li, Yuanfan Zhang, Zuowei Cao, Lei Sun, Panaetov Alexander, Yucong Wang, Minjie Cai, Li Wang, Lu Tian, Zheyuan Wang, Hongbing Ma, Jie Liu, Chao Chen, Yidong Cai, Jie Tang, Gangshan Wu, Weiran Wang, Shirui Huang, Honglei Lu, Huan Liu, Keyan Wang, Jun Chen, Shi Chen, Yuchun Miao, Zimo Huang, Lefei Zhang, Mustafa Ayazoğlu, Wei Xiong, Chengyi Xiong, Fei Wang, Hao Li, Ruimian Wen, Zhijing Yang, Wenbin Zou, Weixin Zheng, Tian Ye, Yuncheng Zhang, Xiangzhen Kong, Aditya Arora, Syed Waqas Zamir, Salman Khan, Munawar Hayat, Fahad Shahbaz Khan, Dandan Gaoand Dengwen Zhouand Qian Ning, Jingzhu Tang, Han Huang, YuFei Wang, Zhangheng Peng, Haobo Li, Wenxue Guan, Shenghua Gong, Xin Li, Jun Liu, Wanjun Wang, Dengwen Zhou, Kun Zeng, Hanjiang Lin, Xinyu Chen, Jinsheng Fang
The aim was to design a network for single image super-resolution that achieved improvement of efficiency measured according to several metrics including runtime, parameters, FLOPs, activations, and memory consumption while at least maintaining the PSNR of 29. 00dB on DIV2K validation set.
no code implementations • 11 Apr 2022 • Gen Li, Laixi Shi, Yuxin Chen, Yuejie Chi, Yuting Wei
We demonstrate that the model-based (or "plug-in") approach achieves minimax-optimal sample complexity without burn-in cost for tabular Markov decision processes (MDPs).
no code implementations • 14 Mar 2022 • Yuling Yan, Gen Li, Yuxin Chen, Jianqing Fan
This paper is concerned with the asynchronous form of Q-learning, which applies a stochastic approximation scheme to Markovian data samples.
no code implementations • 28 Feb 2022 • Laixi Shi, Gen Li, Yuting Wei, Yuxin Chen, Yuejie Chi
Offline or batch reinforcement learning seeks to learn a near-optimal policy using history data without active exploration of the environment.
1 code implementation • 17 Dec 2021 • Jonghyun Kim, Gen Li, Cheolkon Jung, Joongkyu Kim
First, we directly extract the style codes from the original image based on superpixels to consider local objects.
no code implementations • NeurIPS 2021 • Gen Li, Laixi Shi, Yuxin Chen, Yuejie Chi
Achieving sample efficiency in online episodic reinforcement learning (RL) requires optimally balancing exploration and exploitation.
no code implementations • 29 Sep 2021 • Gen Li, Ganghua Wang, Yuantao Gu, Jie Ding
In this paper, the territory of LASSO is extended to the neural network model, a fashionable and powerful nonlinear regression model.
no code implementations • 15 Aug 2021 • Gen Li, Zhen Yang, Yiyong Pan, Jianxiao Ma
It was found that the heavy vehicle takes longer time to complete LC maneuver.
no code implementations • 22 Jun 2021 • Gen Li, Jie Ding
To the best of our knowledge, the rate of convergence of neural networks shown by existing works is bounded by at most the order of $n^{-1/4}$ for a sample size of $n$.
no code implementations • NeurIPS 2021 • Gen Li, Yuxin Chen, Yuejie Chi, Yuantao Gu, Yuting Wei
The current paper pertains to a scenario with value-based linear representation, which postulates the linear realizability of the optimal Q-function (also called the "linear $Q^{\star}$ problem").
no code implementations • 7 Apr 2021 • Gen Li, Changxiao Cai, H. Vincent Poor, Yuxin Chen
Eigenvector perturbation analysis plays a vital role in various data science applications.
2 code implementations • CVPR 2021 • Gen Li, Varun Jampani, Laura Sevilla-Lara, Deqing Sun, Jonghyun Kim, Joongkyu Kim
By integrating the SGC and GPA together, we propose the Adaptive Superpixel-guided Network (ASGNet), which is a lightweight model and adapts to object scale and shape variation.
Ranked #58 on Few-Shot Semantic Segmentation on COCO-20i (5-shot)
no code implementations • 22 Feb 2021 • Gen Li, Yuting Wei, Yuejie Chi, Yuxin Chen
The softmax policy gradient (PG) method, which performs gradient ascent under softmax policy parameterization, is arguably one of the de facto implementations of policy optimization in modern reinforcement learning.
no code implementations • 12 Feb 2021 • Gen Li, Changxiao Cai, Yuxin Chen, Yuting Wei, Yuejie Chi
This paper addresses these questions for the synchronous setting: (1) when $|\mathcal{A}|=1$ (so that Q-learning reduces to TD learning), we prove that the sample complexity of TD learning is minimax optimal and scales as $\frac{|\mathcal{S}|}{(1-\gamma)^3\varepsilon^2}$ (up to log factor); (2) when $|\mathcal{A}|\geq 2$, we settle the sample complexity of Q-learning to be on the order of $\frac{|\mathcal{S}||\mathcal{A}|}{(1-\gamma)^4\varepsilon^2}$ (up to log factor).
no code implementations • 1 Jan 2021 • Gen Li, Yuantao Gu, Jie Ding
A crucial problem in neural networks is to select the most appropriate number of hidden neurons and obtain tight statistical risk bounds.
no code implementations • 2 Oct 2020 • Gen Li, Yuantao Gu, Jie Ding
A crucial problem in neural networks is to select the most appropriate number of hidden neurons and obtain tight statistical risk bounds.
1 code implementation • 27 Aug 2020 • Jonghyun Kim, Gen Li, Inyong Yun, Cheolkon Jung, Joongkyu Kim
In this paper, we propose a novel Edge and Identity Preserving Network for Face SR Network, named as EIPNet, to minimize the distortion by utilizing a lightweight edge block and identity information.
no code implementations • NeurIPS 2020 • Gen Li, Yuting Wei, Yuejie Chi, Yuantao Gu, Yuxin Chen
Focusing on a $\gamma$-discounted MDP with state space $\mathcal{S}$ and action space $\mathcal{A}$, we demonstrate that the $\ell_{\infty}$-based sample complexity of classical asynchronous Q-learning --- namely, the number of samples needed to yield an entrywise $\varepsilon$-accurate estimate of the Q-function --- is at most on the order of $\frac{1}{\mu_{\min}(1-\gamma)^5\varepsilon^2}+ \frac{t_{mix}}{\mu_{\min}(1-\gamma)}$ up to some logarithmic factor, provided that a proper constant learning rate is adopted.
no code implementations • NeurIPS 2020 • Gen Li, Yuting Wei, Yuejie Chi, Yuxin Chen
This paper is concerned with the sample efficiency of reinforcement learning, assuming access to a generative model (or simulator).
Model-based Reinforcement Learning Reinforcement Learning (RL)
no code implementations • NeurIPS 2019 • Changxiao Cai, Gen Li, H. Vincent Poor, Yuxin Chen
We study a noisy tensor completion problem of broad practical interest, namely, the reconstruction of a low-rank tensor from highly incomplete and randomly corrupted observations of its entries.
no code implementations • 9 Oct 2019 • Changxiao Cai, Gen Li, Yuejie Chi, H. Vincent Poor, Yuxin Chen
This paper is concerned with estimating the column space of an unknown low-rank matrix $\boldsymbol{A}^{\star}\in\mathbb{R}^{d_{1}\times d_{2}}$, given noisy and partial observations of its entries.
no code implementations • 16 Aug 2019 • Gen Li, Nan Duan, Yuejian Fang, Ming Gong, Daxin Jiang, Ming Zhou
We propose Unicoder-VL, a universal encoder that aims to learn joint representations of vision and language in a pre-training manner.
Ranked #5 on Image-to-Text Retrieval on MS COCO (Recall@10 metric)
3 code implementations • 26 Jul 2019 • Gen Li, Inyoung Yun, Jonghyun Kim, Joongkyu Kim
As a pixel-level prediction task, semantic segmentation needs large computational cost with enormous parameters to obtain high performance.
no code implementations • 25 Jul 2019 • Gen Li, Yuantao Gu
Spectral Method is a commonly used scheme to cluster data points lying close to Union of Subspaces by first constructing a Random Geometry Graph, called Subspace Clustering.
no code implementations • 14 Jul 2019 • Yuchen Jiao, Gen Li, Yuantao Gu
In this paper, we prove that random projection with the so-called Johnson-Lindenstrauss (JL) property approximately preserves canonical angles between subspaces with overwhelming probability.
no code implementations • 24 May 2019 • Chenfei Wu, Yanzhao Zhou, Gen Li, Nan Duan, Duyu Tang, Xiaojie Wang
This paper presents a strong baseline for real-world visual reasoning (GQA), which achieves 60. 93% in GQA 2019 challenge and won the sixth place.
no code implementations • 23 May 2019 • Xingyu Xv, Gen Li, Yuantao Gu
Subspace Restricted Isometry Property, a newly-proposed concept, has proved to be a useful tool in analyzing the effect of dimensionality reduction algorithms on subspaces.
no code implementations • 3 Oct 2018 • Andrey Ignatov, Radu Timofte, Thang Van Vu, Tung Minh Luu, Trung X. Pham, Cao Van Nguyen, Yongwoo Kim, Jae-Seok Choi, Munchurl Kim, Jie Huang, Jiewen Ran, Chen Xing, Xingguang Zhou, Pengfei Zhu, Mingrui Geng, Yawei Li, Eirikur Agustsson, Shuhang Gu, Luc van Gool, Etienne de Stoutz, Nikolay Kobyshev, Kehui Nie, Yan Zhao, Gen Li, Tong Tong, Qinquan Gao, Liu Hanwen, Pablo Navarrete Michelini, Zhu Dan, Hu Fengshuo, Zheng Hui, Xiumei Wang, Lirui Deng, Rang Meng, Jinghui Qin, Yukai Shi, Wushao Wen, Liang Lin, Ruicheng Feng, Shixiang Wu, Chao Dong, Yu Qiao, Subeesh Vasu, Nimisha Thekke Madam, Praveen Kandula, A. N. Rajagopalan, Jie Liu, Cheolkon Jung
This paper reviews the first challenge on efficient perceptual image enhancement with the focus on deploying deep learning models on smartphones.
1 code implementation • 26 Jul 2018 • Gen Li, Xiaokang Liu, Kun Chen
Multi-view data have been routinely collected in various fields of science and engineering.
no code implementations • 30 Jan 2018 • Gen Li, Qinghua Liu, Yuantao Gu
As an analogy to JL Lemma and RIP for sparse vectors, this work allows the use of random projections to reduce the ambient dimension with the theoretical guarantee that the distance between subspaces after compression is well preserved.
no code implementations • 5 Dec 2017 • Gen Li, Yuchen Jiao, Yuantao Gu
In this work, we study for the first time, without the independence assumption, the convergence behavior of the randomized Kaczmarz method for phase retrieval.
no code implementations • ICCV 2017 • Tong Tong, Gen Li, Xiejie Liu, Qinquan Gao
In this study, we present a novel single-image super-resolution method by introducing dense skip connections in a very deep network.
no code implementations • 16 Aug 2017 • Yanxi Chen, Gen Li, Yuantao Gu
In this letter, we propose a novel Active OMP-SSC, which improves clustering accuracy of OMP-SSC by adaptively updating data points and randomly dropping data points in the OMP process, while still enjoying the low computational complexity of greedy pursuit algorithms.
no code implementations • 20 Jul 2017 • Irina Gaynanova, Gen Li
We call this model SLIDE for Structural Learning and Integrative DEcomposition of multi-view data.
no code implementations • 7 Apr 2017 • Gen Li, Yuantao Gu
Dimension reduction plays an essential role when decreasing the complexity of solving large-scale problems.
no code implementations • 21 Feb 2017 • Yue M. Lu, Gen Li
We study a spectral initialization method that serves a key role in recent work on estimating signals in nonconvex settings.
1 code implementation • 11 Sep 2016 • Eric F. Lock, Gen Li
We describe a likelihood-based latent variable representation of the CP factorization, in which the latent variables are informed by additional covariates.
1 code implementation • 2 Apr 2013 • Susan Wei, Chihoon Lee, Lindsay Wichers, Gen Li, J. S. Marron
Motivated by the prevalence of high dimensional low sample size datasets in modern statistical applications, we propose a general nonparametric framework, Direction-Projection-Permutation (DiProPerm), for testing high dimensional hypotheses.
Methodology