no code implementations • Findings (EMNLP) 2021 • Zixuan Zhang, Hongwei Wang, Han Zhao, Hanghang Tong, Heng Ji
Relations in most of the traditional knowledge graphs (KGs) only reflect static and factual connections, but fail to represent the dynamic activities and state changes about entities.
no code implementations • CONSTRAINT (ACL) 2022 • Ziming Zhou, Han Zhao, Jingjing Dong, Jun Gao, Xiaolong Liu
The memes serve as an important tool in online communication, whereas some hateful memes endanger cyberspace by attacking certain people or subjects.
no code implementations • 25 Apr 2022 • Jing Dong, Shiji Zhou, Baoxiang Wang, Han Zhao
We thus study the problem of supervised gradual domain adaptation, where labeled data from shifting distributions are available to the learner along the trajectory, and we aim to learn a classifier on a target data distribution of interest.
1 code implementation • 18 Apr 2022 • Haoxiang Wang, Bo Li, Han Zhao
Gradual domain adaptation (GDA), on the other hand, assumes a path of $(T-1)$ unlabeled intermediate domains bridging the source and target, and aims to provide better generalization in the target domain by leveraging the intermediate ones.
no code implementations • 9 Mar 2022 • Zhenhailong Wang, Hang Yu, Manling Li, Han Zhao, Heng Ji
Due to the uniform task sampling procedure, MAMF consistently outperforms the classical fine-tuning method for few-shot transfer learning on five benchmark datasets.
1 code implementation • ICLR 2022 • Yao-Hung Hubert Tsai, Tianqin Li, Martin Q. Ma, Han Zhao, Kun Zhang, Louis-Philippe Morency, Ruslan Salakhutdinov
Conditional contrastive learning frameworks consider the conditional sampling procedure that constructs positive or negative data pairs conditioned on specific variables.
1 code implementation • 30 Jan 2022 • Haoxiang Wang, Haozhe Si, Bo Li, Han Zhao
Our first algorithm, ISR-Mean, can identify the subspace spanned by invariant features from the first-order moments of the class-conditional distributions, and achieve provable domain generalization with $d_s+1$ training environments under the data model of Rosenfeld et al. (2021).
no code implementations • 7 Dec 2021 • Han Zhao, Lei Guo
In this paper a novel model-free algorithm is proposed.
1 code implementation • 19 Nov 2021 • Jianfeng Chi, Jian Shen, Xinyi Dai, Weinan Zhang, Yuan Tian, Han Zhao
We first provide a decomposition theorem for return disparity, which decomposes the return disparity of any two MDPs sharing the same state and action spaces into the distance between group-wise reward functions, the discrepancy of group policies, and the discrepancy between state visitation distributions induced by the group policies.
no code implementations • 25 Oct 2021 • Lei Guo, Han Zhao, Yuan Song
First, the deficiency of chattering in traditional SMC and the quasi-SMC method are analyzed in this paper.
1 code implementation • 16 Oct 2021 • Yan Shen, Jian Du, Han Zhao, Benyu Zhang, Zhanghexuan Ji, Mingchen Gao
Federated adversary domain adaptation is a unique distributed minimax training task due to the prevalence of label imbalance among clients, with each client only seeing a subset of the classes of labels required to train a global model.
no code implementations • ICLR 2022 • Ruicheng Xian, Heng Ji, Han Zhao
Recent advances in neural modeling have produced deep multilingual language models capable of extracting cross-lingual knowledge from unparallel texts, as evidenced by their decent zero-shot transfer performance.
no code implementations • 29 Sep 2021 • Xiaoyang Wang, Han Zhao, Klara Nahrstedt, Oluwasanmi O Koyejo
To this end, we propose a strategy to mitigate the effect of spurious features based on our observation that the global model in the federated learning step has a low accuracy disparity due to statistical heterogeneity.
1 code implementation • 16 Jun 2021 • Haoxiang Wang, Han Zhao, Bo Li
Despite the subtle difference between MTL and meta-learning in the problem formulation, both learning paradigms share the same insight that the shared structure between existing training tasks could lead to better generalization and adaptation.
Ranked #13 on
Few-Shot Image Classification
on FC100 5-way (1-shot)
no code implementations • 16 Jun 2021 • Han Zhao
In this paper, we characterize the inherent tradeoff between statistical parity and accuracy in the regression setting by providing a lower bound on the error of any fair regressor.
no code implementations • 11 Jun 2021 • Bo Li, Yifei Shen, Yezhen Wang, Wenzhen Zhu, Colorado J. Reed, Jun Zhang, Dongsheng Li, Kurt Keutzer, Han Zhao
IIB significantly outperforms IRM on synthetic datasets, where the pseudo-invariant features and geometric skews occur, showing the effectiveness of proposed formulation in overcoming failure modes of IRM.
no code implementations • 11 Jun 2021 • Shiji Zhou, Han Zhao, Shanghang Zhang, Lianzhe Wang, Heng Chang, Zhi Wang, Wenwu Zhu
Our theoretical results show that OSAMD can fast adapt to changing environments with active queries.
2 code implementations • NeurIPS 2021 • Guojun Zhang, Han Zhao, YaoLiang Yu, Pascal Poupart
We then prove that our transferability can be estimated with enough samples and give a new upper bound for the target error based on our transferability.
no code implementations • 5 Jun 2021 • Yao-Hung Hubert Tsai, Martin Q. Ma, Han Zhao, Kun Zhang, Louis-Philippe Morency, Ruslan Salakhutdinov
However, data sometimes contains certain information that may be undesirable for downstream tasks.
no code implementations • 19 May 2021 • Lei Guo, Han Zhao
In this paper, we present a novel algorithm named synchronous integral Q-learning, which is based on synchronous policy iteration, to solve the continuous-time infinite horizon optimal control problems of input-affine system dynamics.
no code implementations • 23 Mar 2021 • Xiaolong Chen, Wenyu Liang, Han Zhao, Abdullah Al Mamun
Ultrasonic motors (USMs) are commonly used in aerospace, robotics, and medical devices, where fast and precise motion is needed.
1 code implementation • ICLR 2021 • Yao-Hung Hubert Tsai, Martin Q. Ma, Muqiao Yang, Han Zhao, Louis-Philippe Morency, Ruslan Salakhutdinov
This paper introduces Relative Predictive Coding (RPC), a new contrastive representation learning objective that maintains a good balance among training stability, minibatch size sensitivity, and downstream task performance.
1 code implementation • 24 Feb 2021 • Jianfeng Chi, Yuan Tian, Geoffrey J. Gordon, Han Zhao
With the widespread deployment of large-scale prediction systems in high-stakes domains, e. g., face recognition, criminal justice, etc., disparity in prediction accuracy between different demographic subgroups has called for fundamental understanding on the source of such disparity and algorithmic intervention to mitigate it.
no code implementations • 29 Jan 2021 • Xuecong Sun, Han Jia, Yuzhen Yang, Han Zhao, Yafeng Bi, Zhaoyong Sun, Jun Yang
From ancient to modern times, acoustic structures have been used to control the propagation of acoustic waves.
1 code implementation • ICLR 2021 • Peizhao Li, Yifei Wang, Han Zhao, Pengyu Hong, Hongfu Liu
Disparate impact has raised serious concerns in machine learning applications and its societal impacts.
no code implementations • 19 Dec 2020 • Han Zhao, Chen Dan, Bryon Aragam, Tommi S. Jaakkola, Geoffrey J. Gordon, Pradeep Ravikumar
A wide range of machine learning applications such as privacy-preserving learning, algorithmic fairness, and domain adaptation/generalization among others, involve learning \emph{invariant representations} of the data that aim to achieve two competing goals: (a) maximize information or accuracy with respect to a target response, and (b) maximize invariance or independence with respect to a set of protected features (e. g.\ for fairness, privacy, etc).
1 code implementation • NeurIPS 2020 • Jian Shen, Han Zhao, Weinan Zhang, Yong Yu
However, due to the potential distribution mismatch between simulated data and real data, this could lead to degraded performance.
no code implementations • CVPR 2021 • Bo Li, Yezhen Wang, Shanghang Zhang, Dongsheng Li, Trevor Darrell, Kurt Keutzer, Han Zhao
First, we provide a finite sample bound for both classification and regression problems under Semi-DA.
1 code implementation • 28 Sep 2020 • Peiyuan Liao, Han Zhao, Keyulu Xu, Tommi Jaakkola, Geoffrey Gordon, Stefanie Jegelka, Ruslan Salakhutdinov
While the advent of Graph Neural Networks (GNNs) has greatly improved node and graph representation learning in many applications, the neighborhood aggregation scheme exposes additional vulnerabilities to adversaries seeking to extract node-level information about sensitive attributes.
no code implementations • 15 Sep 2020 • Huajie Shao, Haohong Lin, Qinmin Yang, Shuochao Yao, Han Zhao, Tarek Abdelzaher
Existing methods, such as $\beta$-VAE and FactorVAE, assign a large weight to the KL-divergence term in the objective function, leading to high reconstruction errors for the sake of better disentanglement.
1 code implementation • 1 Sep 2020 • Sicheng Zhao, Xiangyu Yue, Shanghang Zhang, Bo Li, Han Zhao, Bichen Wu, Ravi Krishna, Joseph E. Gonzalez, Alberto L. Sangiovanni-Vincentelli, Sanjit A. Seshia, Kurt Keutzer
To cope with limited labeled training data, many have attempted to directly apply models trained on a large-scale labeled source domain to another sparsely labeled or unlabeled target domain.
no code implementations • ICML 2020 • Han Zhao, Junjie Hu, Andrej Risteski
The goal of universal machine translation is to learn to translate between any pair of languages, given a corpus of paired translated documents for \emph{a small subset} of all pairs of languages.
1 code implementation • NeurIPS 2020 • Yao-Hung Hubert Tsai, Han Zhao, Makoto Yamada, Louis-Philippe Morency, Ruslan Salakhutdinov
Since its inception, the neural estimation of mutual information (MI) has demonstrated the empirical success of modeling expected dependency between high-dimensional random variables.
1 code implementation • NeurIPS 2020 • Remi Tachet, Han Zhao, Yu-Xiang Wang, Geoff Gordon
However, recent work has shown limitations of this approach when label distributions differ between the source and target domains.
no code implementations • ICLR 2020 • Tameem Adel, Han Zhao, Richard E. Turner
Approaches to continual learning aim to successfully learn a set of related tasks that arrive in an online manner.
no code implementations • ICLR 2020 • Han Zhao, Amanda Coston, Tameem Adel, Geoffrey J. Gordon
We propose a novel algorithm for learning fair representations that can simultaneously mitigate two notions of disparity among different demographic subgroups in the classification setting.
no code implementations • 25 Sep 2019 • Han Zhao, Jianfeng Chi, Yuan Tian, Geoffrey J. Gordon
With the prevalence of machine learning services, crowdsourced data containing sensitive information poses substantial privacy challenges.
1 code implementation • NeurIPS 2019 • Han Zhao, Yao-Hung Hubert Tsai, Ruslan Salakhutdinov, Geoffrey J. Gordon
Feed-forward neural networks can be understood as a combination of an intermediate representation and a linear hypothesis.
no code implementations • NeurIPS 2019 • Han Zhao, Geoffrey J. Gordon
On the upside, we prove that if the group-wise Bayes optimal classifiers are close, then learning fair representations leads to an alternative notion of fairness, known as the accuracy parity, which states that the error rates are close between groups.
no code implementations • NeurIPS 2020 • Han Zhao, Jianfeng Chi, Yuan Tian, Geoffrey J. Gordon
Meanwhile, it is clear that in general there is a tension between minimizing information leakage and maximizing task accuracy.
no code implementations • ICLR 2019 • Han Zhao, Yao-Hung Hubert Tsai, Ruslan Salakhutdinov, Geoff Gordon
Learning deep neural networks could be understood as the combination of representation learning and learning halfspaces.
2 code implementations • 27 Jan 2019 • Han Zhao, Remi Tachet des Combes, Kun Zhang, Geoffrey J. Gordon
Our result characterizes a fundamental tradeoff between learning invariant representations and achieving small joint error on both domains when the marginal label distributions differ from source to target.
no code implementations • NeurIPS 2018 • Han Zhao, Shanghang Zhang, Guanhang Wu, José M. F. Moura, Joao P. Costeira, Geoffrey J. Gordon
In this paper we propose new generalization bounds and algorithms under both classification and regression settings for unsupervised multiple source domain adaptation.
1 code implementation • 16 Jun 2018 • Yichong Xu, Han Zhao, Xiaofei Shi, Jeremy Zhang, Nihar B. Shah
We then empirically show that the requisite property on the authorship graph is indeed satisfied in the submission data from the ICLR conference, and further demonstrate a simple trick to make the partitioning method more practically appealing for conference peer review.
no code implementations • 2 May 2018 • Han Zhao, Shuayb Zarar, Ivan Tashev, Chin-Hui Lee
By incorporating prior knowledge of speech signals into the design of model structures, we build a model that is more data-efficient and achieves better generalization on both seen and unseen noise.
no code implementations • 19 Jan 2018 • Chen Liang, Jianbo Ye, Han Zhao, Bart Pursel, C. Lee Giles
Strict partial order is a mathematical structure commonly seen in relational data.
no code implementations • ICLR 2018 • Han Zhao, Shanghang Zhang, Guanhang Wu, Jo\~{a}o P. Costeira, Jos\'{e} M. F. Moura, Geoffrey J. Gordon
We propose a new generalization bound for domain adaptation when there are multiple source domains with labeled instances and one target domain with unlabeled instances.
no code implementations • ICLR 2018 • Yao-Hung Hubert Tsai, Han Zhao, Nebojsa Jojic, Ruslan Salakhutdinov
The assumption that data samples are independently identically distributed is the backbone of many learning algorithms.
no code implementations • ICLR 2018 • Yao-Hung Hubert Tsai, Han Zhao, Ruslan Salakhutdinov, Nebojsa Jojic
In this technical report, we introduce OrderNet that can be used to extract the order of data instances in an unsupervised way.
no code implementations • 20 Jun 2017 • Han Zhao, Geoff Gordon
Symmetric nonnegative matrix factorization has found abundant applications in various domains by providing a symmetric low-rank decomposition of nonnegative matrices.
3 code implementations • 26 May 2017 • Han Zhao, Shanghang Zhang, Guanhang Wu, João P. Costeira, José M. F. Moura, Geoffrey J. Gordon
As a step toward bridging the gap, we propose a new generalization bound for domain adaptation when there are multiple source domains with labeled instances and one target domain with unlabeled instances.
no code implementations • ICLR 2018 • Han Zhao, Zhenyao Zhu, Junjie Hu, Adam Coates, Geoff Gordon
This provides us a very general way to interpolate between generative and discriminative extremes through different choices of priors.
no code implementations • NeurIPS 2017 • Han Zhao, Geoff Gordon
We propose a dynamic programming method to further reduce the computation of the moments of all the edges in the graph from quadratic to linear.
no code implementations • 14 Feb 2017 • Han Zhao, Otilia Stretcu, Alex Smola, Geoff Gordon
In this paper, we consider a formulation of multitask learning that learns the relationships both between tasks and between features, represented through a task covariance and a feature covariance matrix, respectively.
no code implementations • NeurIPS 2016 • Han Zhao, Pascal Poupart, Geoff Gordon
We present a unified approach for learning the parameters of Sum-Product networks (SPNs).
1 code implementation • 20 Apr 2015 • Han Zhao, Zhengdong Lu, Pascal Poupart
The ability to accurately model a sentence at varying stages (e. g., word-phrase-sentence) plays a central role in natural language processing.
Ranked #5 on
Subjectivity Analysis
on SUBJ
no code implementations • 6 Jan 2015 • Han Zhao, Mazen Melibari, Pascal Poupart
We conclude the paper with some discussion of the implications of the proof and establish a connection between the depth of an SPN and a lower bound of the tree-width of its corresponding BN.
no code implementations • 18 Jun 2014 • Han Zhao, Pascal Poupart
In contrast, maximum likelihood estimates may get trapped in local optima due to the non-convex nature of the likelihood function of latent variable models.