1 code implementation • ICML 2020 • Kaiwen Wu, Allen Houze Wang, Yao-Liang Yu
While the majority of existing attacks focus on measuring perturbations under the $\ell_p$ metric, Wasserstein distance, which takes geometry in pixel space into account, has long been known to be a suitable metric for measuring image quality and has recently risen as a compelling alternative to the $\ell_p$ metric in adversarial attacks.
1 code implementation • 25 Jun 2020 • Guojun Zhang, Kaiwen Wu, Pascal Poupart, Yao-Liang Yu
We prove their local convergence at strict local minimax points, which are surrogates of global solutions.
2 code implementations • 20 Jun 2020 • Zeou Hu, Kiarash Shaloudegi, Guojun Zhang, Yao-Liang Yu
Federated learning has emerged as a promising, massively distributed way to train a joint deep model over large amounts of edge devices while keeping private user data strictly on device.
1 code implementation • 16 Jun 2020 • Tim Dockhorn, James A. Ritchie, Yao-Liang Yu, Iain Murray
Density deconvolution is the task of estimating a probability density function given only noise-corrupted samples.
3 code implementations • 25 May 2020 • Takanori Fujiwara, Jian Zhao, Francine Chen, Yao-Liang Yu, Kwan-Liu Ma
This analysis task could be greatly assisted by contrastive learning, which is an emerging analysis approach to discover salient patterns in one dataset relative to another.
1 code implementation • ACL 2020 • Raphael Tang, Jaejun Lee, Ji Xin, Xinyu Liu, Yao-Liang Yu, Jimmy Lin
In natural language processing, a recently popular line of work explores how to best report the experimental results of neural networks.
3 code implementations • ACL 2020 • Ji Xin, Raphael Tang, Jaejun Lee, Yao-Liang Yu, Jimmy Lin
Large-scale pre-trained language models such as BERT have brought significant improvements to NLP applications.
no code implementations • ICML 2020 • Yingyi Ma, Vignesh Ganapathiraman, Yao-Liang Yu, Xinhua Zhang
Invariance (defined in a general sense) has been one of the most effective priors for representation learning.
no code implementations • 8 Mar 2020 • Allen Houze Wang, Priyank Jaini, Yao-Liang Yu, Pascal Poupart
Recently, the conditional SAGE certificate has been proposed as a sufficient condition for signomial positivity over a convex set.
no code implementations • 27 Feb 2020 • Guojun Zhang, Pascal Poupart, Yao-Liang Yu
Convergence to a saddle point for convex-concave functions has been studied for decades, while recent years has seen a surge of interest in non-convex (zero-sum) smooth games, motivated by their recent wide applications.
1 code implementation • 28 Jan 2020 • Xin Lian, Kshitij Jain, Jakub Truszkowski, Pascal Poupart, Yao-Liang Yu
We study unsupervised multilingual alignment, the problem of finding word-to-word translations between multiple languages without using any parallel data.
Ranked #1 on Word Alignment on en-es
1 code implementation • NeurIPS 2019 • Jingjing Wang, Sun Sun, Yao-Liang Yu
Novelty detection, a fundamental task in machine learning, has drawn a lot of recent attention due to its wide-ranging applications and the rise of neural approaches.
no code implementations • 15 Nov 2019 • Achyudh Ram, Ji Xin, Meiyappan Nagappan, Yao-Liang Yu, Rocío Cabrera Lozoya, Antonino Sabetta, Jimmy Lin
Public vulnerability databases such as CVE and NVD account for only 60% of security vulnerabilities present in open-source projects, and are known to suffer from inconsistent quality.
no code implementations • IJCNLP 2019 • Ji Xin, Jimmy Lin, Yao-Liang Yu
Memory neurons of long short-term memory (LSTM) networks encode and process information in powerful yet mysterious ways.
1 code implementation • ICLR 2020 • Guojun Zhang, Yao-Liang Yu
Min-max formulations have attracted great attention in the ML community due to the rise of deep generative models and adversarial methods, while understanding the dynamics of gradient algorithms for solving such formulations has remained a grand challenge.
no code implementations • 26 Jul 2019 • Kaiwen Wu, Yao-Liang Yu
Deep models, while being extremely versatile and accurate, are vulnerable to adversarial attacks: slight perturbations that are imperceptible to humans can completely flip the prediction of deep models.
no code implementations • ICML 2020 • Priyank Jaini, Ivan Kobyzev, Yao-Liang Yu, Marcus Brubaker
We investigate the ability of popular flow based methods to capture tail-properties of a target density by studying the increasing triangular maps used in these flow methods acting on a tractable source density.
no code implementations • 13 May 2019 • Borislav Mavrin, Shangtong Zhang, Hengshuai Yao, Linglong Kong, Kaiwen Wu, Yao-Liang Yu
In distributional reinforcement learning (RL), the estimated distribution of value function models both the parametric and intrinsic uncertainties.
2 code implementations • 7 May 2019 • Priyank Jaini, Kira A. Selby, Yao-Liang Yu
Triangular map is a recent construct in probability theory that allows one to transform any source probability density function to any target density function.
no code implementations • NeurIPS 2018 • Priyank Jaini, Pascal Poupart, Yao-Liang Yu
At their core, many unsupervised learning models provide a compact representation of homogeneous density mixtures, but their similarities and differences are not always clearly understood.
no code implementations • ICML 2018 • Vignesh Ganapathiraman, Zhan Shi, Xinhua Zhang, Yao-Liang Yu
Latent prediction models, exemplified by multi-layer networks, employ hidden variables that automate abstract feature discovery.
1 code implementation • 6 Mar 2018 • Seojin Bang, Yao-Liang Yu, Wei Wu
To address this problem and inspired by recent works in adversarial learning, we propose a multiple kernel clustering method with the min-max framework that aims to be robust to such adversarial perturbation.
no code implementations • NeurIPS 2017 • Zhan Shi, Xinhua Zhang, Yao-Liang Yu
Adversarial machines, where a learner competes against an adversary, have regained much recent interest in machine learning.
no code implementations • 30 Nov 2017 • Shrinu Kushagra, Yao-Liang Yu, Shai Ben-David
We focus on the $k$-means objective and we prove that the regularised version of $k$-means is NP-Hard even for $k=1$.
no code implementations • ICML 2017 • Pengtao Xie, Yuntian Deng, Yi Zhou, Abhimanu Kumar, Yao-Liang Yu, James Zou, Eric P. Xing
The large model capacity of latent space models (LSMs) enables them to achieve great performance on various applications, but meanwhile renders LSMs to be prone to overfitting.
no code implementations • CVPR 2017 • Marc T. Law, Yao-Liang Yu, Raquel Urtasun, Richard S. Zemel, Eric P. Xing
Classic approaches alternate the optimization over the learned metric and the assignment of similar instances.
no code implementations • 1 May 2017 • Junming Yin, Yao-Liang Yu
Sparse additive modeling is a class of effective methods for performing high-dimensional nonparametric regression.
no code implementations • NeurIPS 2016 • Vignesh Ganapathiraman, Xinhua Zhang, Yao-Liang Yu, Junfeng Wen
Unsupervised learning of structured predictors has been a long standing pursuit in machine learning.
no code implementations • 26 Sep 2016 • Xuezhe Ma, Yingkai Gao, Zhiting Hu, Yao-Liang Yu, Yuntian Deng, Eduard Hovy
Algorithmically, we show that our proposed measure of the inference gap can be used to regularize the standard dropout training objective, resulting in an \emph{explicit} control of the gap.
no code implementations • CVPR 2016 • Xiaojun Chang, Yao-Liang Yu, Yi Yang, Eric P. Xing
Complex event detection on unconstrained Internet videos has seen much progress in recent years.
no code implementations • CVPR 2016 • Marc T. Law, Yao-Liang Yu, Matthieu Cord, Eric P. Xing
Clustering is the task of grouping a set of objects so that objects in the same cluster are more similar to each other than to those in other clusters.
2 code implementations • 31 Jan 2016 • Kirthevasan Kandasamy, Yao-Liang Yu
Between non-additive models which often have large variance and first order additive models which have large bias, there has been little work to exploit the trade-off in the middle via additive models of intermediate order.
no code implementations • 26 Nov 2015 • Pengtao Xie, Jin Kyu Kim, Yi Zhou, Qirong Ho, Abhimanu Kumar, Yao-Liang Yu, Eric Xing
Matrix-parametrized models, including multiclass logistic regression and sparse coding, are used in machine learning (ML) applications ranging from computer vision to computational biology.
no code implementations • 17 Oct 2014 • Yao-Liang Yu, Xinhua Zhang, Dale Schuurmans
Structured sparsity is an important modeling tool that expands the applicability of convex formulations for data analysis, however it also creates significant challenges for efficient algorithm design.
no code implementations • 19 Sep 2014 • Pengtao Xie, Jin Kyu Kim, Yi Zhou, Qirong Ho, Abhimanu Kumar, Yao-Liang Yu, Eric Xing
Matrix-parametrized models, including multiclass logistic regression and sparse coding, are used in machine learning (ML) applications ranging from computer vision to computational biology.
no code implementations • 30 Dec 2013 • Eric P. Xing, Qirong Ho, Wei Dai, Jin Kyu Kim, Jinliang Wei, Seunghak Lee, Xun Zheng, Pengtao Xie, Abhimanu Kumar, Yao-Liang Yu
What is a systematic way to efficiently apply a wide spectrum of advanced ML programs to industrial scale problems, using Big Models (up to 100s of billions of parameters) on Big Data (up to terabytes or petabytes)?
no code implementations • NeurIPS 2013 • Xinhua Zhang, Yao-Liang Yu, Dale Schuurmans
Structured sparse estimation has become an important technique in many areas of data analysis.
no code implementations • NeurIPS 2013 • Yao-Liang Yu
It is a common practice to approximate complicated'' functions with more friendly ones.
no code implementations • NeurIPS 2013 • Yao-Liang Yu
The proximal map is the key step in gradient-type algorithms, which have become prevalent in large-scale high-dimensional problems.
no code implementations • NeurIPS 2012 • Yao-Liang Yu, Özlem Aslan, Dale Schuurmans
Despite the variety of robust regression methods that have been developed, current regression formulations are either NP-hard, or allow unbounded response to even a single leverage point.
no code implementations • NeurIPS 2012 • Martha White, Xinhua Zhang, Dale Schuurmans, Yao-Liang Yu
Subspace learning seeks a low dimensional representation of data that enables accurate reconstruction.
no code implementations • NeurIPS 2012 • Xinhua Zhang, Dale Schuurmans, Yao-Liang Yu
Sparse learning models typically combine a smooth loss with a nonsmooth penalty, such as trace norm.
no code implementations • NeurIPS 2010 • Min Yang, Linli Xu, Martha White, Dale Schuurmans, Yao-Liang Yu
We present a generic procedure that can be applied to standard loss functions and demonstrate improved robustness in regression and classification problems.
no code implementations • NeurIPS 2009 • Yao-Liang Yu, Yuxi Li, Dale Schuurmans, Csaba Szepesvári
We prove that linear projections between distribution families with fixed first and second moments are surjective, regardless of dimension.