no code implementations • ICML 2018 • Rie Johnson, Tong Zhang
This paper first presents a theory for generative adversarial methods that does not rely on the traditional minimax formulation.
no code implementations • ICML 2018 • Weizhong Zhang, Bin Hong, Lin Ma, Wei Liu, Tong Zhang
Relying on this study, we subsequently propose a novel safe screening method to quickly identify the elements guaranteed to be included (we refer to them as active) or excluded (inactive) in the final optimal solution of SFM during the optimization process.
no code implementations • ICML 2018 • Wenhan Luo, Peng Sun, Fangwei Zhong, Wei Liu, Tong Zhang, Yizhou Wang
We study active object tracking, where a tracker takes as input the visual observation (i. e., frame sequence) and produces the camera control signal (e. g., move forward, turn left, etc.).
no code implementations • NeurIPS 2018 • Hanlin Tang, Shaoduo Gan, Ce Zhang, Tong Zhang, Ji Liu
In this paper, We explore a natural question: {\em can the combination of both techniques lead to a system that is robust to both bandwidth and latency?}
1 code implementation • EMNLP 2018 • Yi Liao, Lidong Bing, Piji Li, Shuming Shi, Wai Lam, Tong Zhang
For example, an input sequence could be a word sequence, such as review sentence and advertisement text.
no code implementations • 16 Apr 2018 • Jiatao Jiang, Chunyan Xu, Zhen Cui, Tong Zhang, Wenming Zheng, Jian Yang
As an analogy to a standard convolution kernel on image, Gaussian models implicitly coordinate those unordered vertices/nodes and edges in a local receptive field after projecting to the gradient space of Gaussian parameters.
no code implementations • 4 Apr 2018 • Xinpeng Chen, Jingyuan Chen, Lin Ma, Jian Yao, Wei Liu, Jiebo Luo, Tong Zhang
First, we demonstrate that video attractiveness and different engagements present different relationships.
no code implementations • CVPR 2018 • Jing Zhang, Tong Zhang, Yuchao Dai, Mehrtash Harandi, Richard Hartley
Such supervision, while labor-intensive and not always possible, tends to hinder the generalization ability of the learned models.
no code implementations • 27 Mar 2018 • Tong Zhang, Wenming Zheng, Zhen Cui, Yang Li
For cross graph convolution, a parameterized Kronecker sum operation is proposed to generate a conjunctive adjacency matrix characterizing the relationship between every pair of nodes across two subgraphs.
no code implementations • ECCV 2018 • Xinyu Gong, HaoZhi Huang, Lin Ma, Fumin Shen, Wei Liu, Tong Zhang
While each view of the stereoscopic pair is processed in an individual path, a novel feature aggregation strategy is proposed to effectively share information between the two paths.
no code implementations • 19 Jun 2017 • Xingguo Li, Lin F. Yang, Jason Ge, Jarvis Haupt, Tong Zhang, Tuo Zhao
We propose a DC proximal Newton algorithm for solving nonconvex regularized sparse learning problems in high dimensions.
no code implementations • 15 May 2017 • Luo Luo, Cheng Chen, Zhihua Zhang, Wu-Jun Li, Tong Zhang
We also apply RFD to online learning and propose an effective hyperparameter-free online Newton algorithm.
no code implementations • ICML 2018 • Lei Han, Yiheng Huang, Tong Zhang
This paper proposes a method for multi-class classification problems, where the number of classes K is large.
no code implementations • NeurIPS 2018 • Jianqiao Wangni, Jialei Wang, Ji Liu, Tong Zhang
Modern large scale machine learning applications require stochastic optimization algorithms to be implemented on distributed computational architectures.
no code implementations • 21 Jun 2017 • Jialei Wang, Tong Zhang
We present novel minibatch stochastic optimization methods for empirical risk minimization problems, the methods efficiently leverage variance reduced first-order and sub-sampled higher-order information to accelerate the convergence speed.
no code implementations • 16 Mar 2016 • Chris Junchi Li, Mengdi Wang, Han Liu, Tong Zhang
We prove for the first time a nearly optimal finite-sample error bound for the online PCA algorithm.
no code implementations • 30 May 2017 • Tong Zhang, Wenming Zheng, Zhen Cui, Chaolong Li
Symmetric positive definite (SPD) matrices (e. g., covariances, graph Laplacians, etc.)
no code implementations • 13 Apr 2016 • Shun Zheng, Jialei Wang, Fen Xia, Wei Xu, Tong Zhang
In modern large-scale machine learning applications, the training data are often partitioned and stored on multiple machines.
no code implementations • 4 Jun 2017 • Qiang Sun, Kean Ming Tan, Han Liu, Tong Zhang
Our proposal is computationally tractable and produces an estimator that achieves the oracle rate of convergence.
no code implementations • 12 May 2017 • Tong Zhang, Wenming Zheng, Zhen Cui, Yuan Zong, Yang Li
Then a bi-directional temporal RNN layer is further used to learn discriminative temporal dependencies from the sequences concatenating spatial features of each time slice produced from the spatial RNN layer.
no code implementations • 23 Dec 2014 • Tuo Zhao, Han Liu, Tong Zhang
This is the first result on the computational and statistical guarantees of the pathwise coordinate optimization framework in high dimensions.
no code implementations • 29 Mar 2015 • Shusen Wang, Zhihua Zhang, Tong Zhang
The Nystr\"om method is a special instance of our fast model and is approximation to the prototype model.
no code implementations • 31 Aug 2016 • Rie Johnson, Tong Zhang
This paper reports the performances of shallow word-level convolutional neural networks (CNN), our earlier work (2015), on the eight datasets with relatively large training data that were used for testing the very deep character-level CNN in Conneau et al. (2016).
no code implementations • 27 Apr 2016 • Lei Han, Kean Ming Tan, Ting Yang, Tong Zhang
A major challenge for building statistical models in the big data era is that the available data volume far exceeds the computational capability.
no code implementations • 7 Feb 2016 • Rie Johnson, Tong Zhang
The best results were obtained by combining region embeddings in the form of LSTM and convolution layers trained on unlabeled data.
Ranked #1 on Text Classification on RCV1
no code implementations • ICML 2017 • Jialei Wang, Mladen Kolar, Nathan Srebro, Tong Zhang
We propose a novel, efficient approach for distributed sparse learning in high-dimensions, where observations are randomly partitioned across machines.
no code implementations • 29 Apr 2016 • Kean Ming Tan, Zhaoran Wang, Han Liu, Tong Zhang
Sparse generalized eigenvalue problem (GEP) plays a pivotal role in a large family of high-dimensional statistical models, including sparse Fisher's discriminant analysis, canonical correlation analysis, and sufficient dimension reduction.
no code implementations • 13 Dec 2012 • Sivan Sabato, Shai Shalev-Shwartz, Nathan Srebro, Daniel Hsu, Tong Zhang
We consider the problem of learning a non-negative linear classifier with a $1$-norm of at most $k$, and a fixed threshold, under the hinge-loss.
no code implementations • 14 Nov 2015 • Zhuoran Yang, Zhaoran Wang, Han Liu, Yonina C. Eldar, Tong Zhang
To recover $\beta^*$, we propose an $\ell_1$-regularized least-squares estimator.
no code implementations • NeurIPS 2015 • Rie Johnson, Tong Zhang
This paper presents a new semi-supervised framework with convolutional neural networks (CNNs) for text categorization.
Ranked #1000000000 on Text Classification on IMDb
no code implementations • 26 Dec 2014 • Shusen Wang, Tong Zhang, Zhihua Zhang
Low-rank matrix completion is an important problem with extensive real-world applications.
no code implementations • 20 Jun 2013 • Zhaoran Wang, Han Liu, Tong Zhang
In particular, our analysis improves upon existing results by providing a more refined sample complexity bound as well as an exact support recovery result for the final estimator.
no code implementations • 13 Jan 2014 • Peilin Zhao, Tong Zhang
Uniform sampling of training data has been commonly used in traditional stochastic optimization algorithms such as Proximal Stochastic Gradient Descent (prox-SGD) and Proximal Stochastic Dual Coordinate Ascent (prox-SDCA).
no code implementations • 21 Nov 2014 • Zheng Qu, Peter Richtárik, Tong Zhang
The distributed variant of Quartz is the first distributed SDCA-like method with an analysis for non-separable data.
no code implementations • 16 Dec 2013 • Peilin Zhao, Jinwei Yang, Tong Zhang, Ping Li
The Alternating Direction Method of Multipliers (ADMM) has been studied for years.
no code implementations • 13 May 2014 • Peilin Zhao, Tong Zhang
Stochastic Gradient Descent (SGD) is a popular optimization method which has been applied to many important machine learning tasks such as Support Vector Machines and Deep Neural Networks.
no code implementations • 13 Jun 2011 • Daniel Hsu, Sham M. Kakade, Tong Zhang
The analysis also reveals the effect of errors in the estimated covariance structure, as well as the effect of modeling errors, neither of which effects are present in the fixed design setting.
no code implementations • 19 Mar 2014 • Lin Xiao, Tong Zhang
We consider the problem of minimizing the sum of two convex functions: one is the average of a large number of smooth component functions, and the other is a general convex function that admits a simple proximal mapping.
no code implementations • 31 Dec 2013 • Ping Li, Cun-Hui Zhang, Tong Zhang
In this paper, we adopt very sparse Compressed Counting for nonnegative signal recovery.
no code implementations • 22 Nov 2013 • Xiao-Tong Yuan, Ping Li, Tong Zhang
Numerical evidences show that our method is superior to the state-of-the-art greedy selection methods in sparse logistic regression and sparse precision matrix estimation tasks.
no code implementations • 21 Nov 2013 • Xiao-Tong Yuan, Ping Li, Tong Zhang
We investigate a generic problem of learning pairwise exponential family graphical models with pairwise sufficient statistics defined by a global mapping function, e. g., Mercer kernels.
no code implementations • 12 Nov 2013 • Dong Dai, Philippe Rigollet, Lucy Xia, Tong Zhang
While results indicate that the same aggregation scheme may not satisfy sharp oracle inequalities with high probability, we prove that a weaker notion of oracle inequality for EW that holds with high probability.
no code implementations • 10 Sep 2013 • Shai Shalev-Shwartz, Tong Zhang
We introduce a proximal version of the stochastic dual coordinate ascent method and show how to accelerate the method using an inner-outer iteration procedure.
no code implementations • 3 Oct 2013 • Ping Li, Cun-Hui Zhang, Tong Zhang
In particular, when p->0 the required number of measurements is essentially M=K\log N, where K is the number of nonzero coordinates of the signal.
no code implementations • 26 Sep 2013 • Krishnakumar Balasubramanian, Kai Yu, Tong Zhang
The traditional convex formulation employs the group Lasso relaxation to achieve joint sparsity across tasks.
no code implementations • NeurIPS 2013 • Shai Shalev-Shwartz, Tong Zhang
Stochastic dual coordinate ascent (SDCA) is an effective technique for solving regularized loss minimization problems in machine learning.
no code implementations • ICML 2018 • Jiaxiang Wu, Weidong Huang, Junzhou Huang, Tong Zhang
Large-scale distributed optimization is of great importance in various applications.
no code implementations • NeurIPS 2018 • Cong Fang, Chris Junchi Li, Zhouchen Lin, Tong Zhang
For stochastic first-order method, combining SPIDER with normalized gradient descent, we propose two new algorithms, namely SPIDER-SFO and SPIDER-SFO\textsuperscript{+}, that solve non-convex stochastic optimization problems using stochastic gradients only.
no code implementations • 7 Jul 2018 • Wenting Zhao, Chunyan Xu, Zhen Cui, Tong Zhang, Jiatao Jiang, Zhen-Yu Zhang, Jian Yang
In this paper, we aim to give a comprehensive analysis of when work matters by transforming different classical network structures to graph CNN, particularly in the basic graph recognition problem.
Ranked #3 on Graph Classification on IMDb-B
no code implementations • ECCV 2018 • Minjun Li, Hao-Zhi Huang, Lin Ma, Wei Liu, Tong Zhang, Yu-Gang Jiang
Recent studies on unsupervised image-to-image translation have made a remarkable progress by training a pair of generative adversarial networks with a cycle-consistent loss.
no code implementations • ECCV 2018 • Wenhao Jiang, Lin Ma, Yu-Gang Jiang, Wei Liu, Tong Zhang
In this paper, in order to exploit the complementary information from multiple encoders, we propose a novel Recurrent Fusion Network (RFNet) for tackling image captioning.
no code implementations • 10 Aug 2018 • Wenhan Luo, Peng Sun, Fangwei Zhong, Wei Liu, Tong Zhang, Yizhou Wang
We further propose an environment augmentation technique and a customized reward function, which are crucial for successful training.
no code implementations • NeurIPS 2017 • Chris Junchi Li, Mengdi Wang, Han Liu, Tong Zhang
In this paper, we propose to adopt the diffusion approximation tools to study the dynamics of Oja's iteration which is an online stochastic gradient descent method for the principal component analysis.
no code implementations • 17 Sep 2018 • Kean Ming Tan, Zhaoran Wang, Tong Zhang, Han Liu, R. Dennis Cook
Sliced inverse regression is a popular tool for sufficient dimension reduction, which replaces covariates with a minimal set of their linear combinations without loss of information on the conditional distribution of the response given the covariates.
no code implementations • 25 Sep 2018 • Chaobing Song, Ji Liu, Han Liu, Yong Jiang, Tong Zhang
Regularized online learning is widely used in machine learning applications.
no code implementations • ECCV 2018 • Yonggen Ling, Linchao Bao, Zequn Jie, Fengming Zhu, Ziyang Li, Shanmin Tang, Yongsheng Liu, Wei Liu, Tong Zhang
Our approach is able to handle the rolling-shutter effects and imperfect sensor synchronization in a unified way.
no code implementations • ECCV 2018 • Yitong Wang, Dihong Gong, Zheng Zhou, Xing Ji, Hao Wang, Zhifeng Li, Wei Liu, Tong Zhang
Extensive experiments conducted on the three public domain face aging datasets (MORPH Album 2, CACD-VS and FG-NET) have shown the effectiveness of the proposed approach and the value of the constructed CAF dataset on AIFR.
Ranked #3 on Age-Invariant Face Recognition on MORPH Album2
no code implementations • EMNLP 2018 • Jian Li, Zhaopeng Tu, Baosong Yang, Michael R. Lyu, Tong Zhang
Multi-head attention is appealing for the ability to jointly attend to information from different representation subspaces at different positions.
no code implementations • EMNLP 2018 • Baosong Yang, Zhaopeng Tu, Derek F. Wong, Fandong Meng, Lidia S. Chao, Tong Zhang
Self-attention networks have proven to be of profound value for its strength of capturing global dependencies.
Ranked #29 on Machine Translation on WMT2014 English-German
no code implementations • EMNLP 2018 • Zi-Yi Dou, Zhaopeng Tu, Xing Wang, Shuming Shi, Tong Zhang
Advanced neural machine translation (NMT) models generally implement encoder and decoder as multiple layers, which allows systems to model complex functions and capture complicated linguistic structures.
no code implementations • 2 Nov 2018 • Shixiang Chen, Shiqian Ma, Anthony Man-Cho So, Tong Zhang
We prove that the proposed method globally converges to a stationary point.
no code implementations • NeurIPS 2018 • Conghui Tan, Tong Zhang, Shiqian Ma, Ji Liu
Regularized empirical risk minimization problem with linear predictor appears frequently in machine learning.
no code implementations • 2 Nov 2018 • Tong Zhang, Pan Ji, Mehrtash Harandi, Richard Hartley, Ian Reid
In this paper, we introduce a method that simultaneously learns an embedding space along subspaces within it to minimize a notion of reconstruction error, thus addressing the problem of subspace clustering in an end-to-end learning paradigm.
no code implementations • ECCV 2018 • Kaipeng Zhang, Zhanpeng Zhang, Chia-Wen Cheng, Winston H. Hsu, Yu Qiao, Wei Liu, Tong Zhang
Face hallucination is a generative task to super-resolve the facial image with low resolution while human perception of face heavily relies on identity information.
no code implementations • 21 Nov 2018 • Xiang Kong, Zhaopeng Tu, Shuming Shi, Eduard Hovy, Tong Zhang
Although Neural Machine Translation (NMT) models have advanced state-of-the-art performance in machine translation, they face problems like the inadequate translation.
Ranked #35 on Machine Translation on WMT2014 English-German
no code implementations • 30 Nov 2018 • Keyu Yan, Wenming Zheng, Tong Zhang, Yuan Zong, Zhen Cui
Cross-database non-frontal expression recognition is a very meaningful but rather difficult subject in the fields of computer vision and affect computing.
Facial Expression Recognition Facial Expression Recognition (FER) +1
no code implementations • 6 Dec 2018 • Kaiqing Zhang, Zhuoran Yang, Han Liu, Tong Zhang, Tamer Başar
This work appears to be the first finite-sample analysis for batch MARL, a step towards rigorous theoretical understanding of general MARL algorithms in the finite-sample regime.
Multi-agent Reinforcement Learning reinforcement-learning +1
no code implementations • 19 Dec 2018 • Yuan Zong, Tong Zhang, Wenming Zheng, Xiaopeng Hong, Chuangao Tang, Zhen Cui, Guoying Zhao
Cross-database micro-expression recognition (CDMER) is one of recently emerging and interesting problem in micro-expression analysis.
1 code implementation • EMNLP 2018 • Yi Liao, Lidong Bing, Piji Li, Shuming Shi, Wai Lam, Tong Zhang
For example, an input sequence could be a word sequence, such as review sentence and advertisement text.
no code implementations • NeurIPS 2018 • Cong Fang, Chris Junchi Li, Zhouchen Lin, Tong Zhang
Specially, we prove that the SPIDER-SFO algorithm achieves a gradient computation cost of $\mathcal{O}\left( \min( n^{1/2} \epsilon^{-2}, \epsilon^{-3} ) \right)$ to find an $\epsilon$-approximate first-order stationary point.
no code implementations • NeurIPS 2018 • Conghui Tan, Tong Zhang, Shiqian Ma, Ji Liu
Regularized empirical risk minimization problem with linear predictor appears frequently in machine learning.
no code implementations • NeurIPS 2017 • Xingguo Li, Lin Yang, Jason Ge, Jarvis Haupt, Tong Zhang, Tuo Zhao
We propose a DC proximal Newton algorithm for solving nonconvex regularized sparse learning problems in high dimensions.
no code implementations • NeurIPS 2017 • Wenbing Huang, Mehrtash Harandi, Tong Zhang, Lijie Fan, Fuchun Sun, Junzhou Huang
Linear Dynamical Systems (LDSs) are fundamental tools for modeling spatio-temporal data in various disciplines.
no code implementations • NeurIPS 2016 • Xiaotong Yuan, Ping Li, Tong Zhang
In this paper, we bridge this gap by showing, for the first time, that exact recovery of the global sparse minimizer is possible for HTP-style methods under restricted strong condition number bounding conditions.
no code implementations • NeurIPS 2016 • Xiaotong Yuan, Ping Li, Tong Zhang, Qingshan Liu, Guangcan Liu
We investigate a subclass of exponential family graphical models of which the sufficient statistics are defined by arbitrary additive forms.
no code implementations • NeurIPS 2015 • Zheng Qu, Peter Richtarik, Tong Zhang
We study the problem of minimizing the average of a large number of smooth convex functions penalized with a strongly convex regularizer.
no code implementations • NeurIPS 2015 • Daniel Vainsencher, Han Liu, Tong Zhang
Abstract We propose a family of non-uniform sampling strategies to provably speed up a class of stochastic optimization algorithms with linear convergence including Stochastic Variance Reduced Gradient (SVRG) and Stochastic Dual Coordinate Ascent (SDCA).
no code implementations • NeurIPS 2013 • Rie Johnson, Tong Zhang
Stochastic gradient descent is popular for large scale optimization but has slow convergence asymptotically due to the inherent variance.
no code implementations • NeurIPS 2012 • Quanquan Gu, Tong Zhang, Jiawei Han, Chris H. Ding
In particular, we derive a deterministic generalization error bound for LapRLS trained on subsampled data, and propose to select a subset of data points to label by minimizing this upper bound.
no code implementations • NeurIPS 2011 • Dong Dai, Tong Zhang
The purpose of this paper is to present a new greedy model averaging procedure that improves EWMA.
no code implementations • NeurIPS 2011 • Zhen Li, Huazhong Ning, Liangliang Cao, Tong Zhang, Yihong Gong, Thomas S. Huang
Traditional approaches relied on algorithmic constructions that are often data independent (such as Locality Sensitive Hashing) or weakly dependent (such as kd-trees, k-means trees).
no code implementations • NeurIPS 2011 • Animashree Anandkumar, Kamalika Chaudhuri, Daniel J. Hsu, Sham M. Kakade, Le Song, Tong Zhang
The setting is one where we only have samples from certain observed variables in the tree, and our goal is to estimate the tree structure (i. e., the graph of how the underlying hidden variables are connected to each other and to the observed variables).
no code implementations • NeurIPS 2010 • Alina Beygelzimer, Daniel J. Hsu, John Langford, Tong Zhang
We present and analyze an agnostic active learning algorithm that works without keeping a version space.
no code implementations • NeurIPS 2010 • Yuanqing Lin, Tong Zhang, Shenghuo Zhu, Kai Yu
This paper proposes a principled extension of the traditional single-layer flat sparse coding scheme, where a two-layer coding scheme is derived based on theoretical analysis of nonlinear functional approximation that extends recent results for local coordinate coding.
no code implementations • NeurIPS 2009 • Kai Yu, Tong Zhang, Yihong Gong
This paper introduces a new method for semi-supervised learning on high dimensional nonlinear manifolds, which includes a phase of unsupervised basis learning and a phase of supervised function learning.
no code implementations • NeurIPS 2008 • John Langford, Lihong Li, Tong Zhang
We propose a general method called truncated gradient to induce sparsity in the weights of online-learning algorithms with convex loss.
no code implementations • NeurIPS 2008 • Tong Zhang
Consider linear prediction models where the target function is a sparse linear combination of a set of basis functions.
no code implementations • NeurIPS 2008 • Tong Zhang
We study learning formulations with non-convex regularizaton that are natural for sparse linear models.
no code implementations • NeurIPS 2007 • John Langford, Tong Zhang
We present Epoch-Greedy, an algorithm for multi-armed bandits with observable side information.
no code implementations • NeurIPS 2007 • Zhaohui Zheng, Hongyuan Zha, Tong Zhang, Olivier Chapelle, Keke Chen, Gordon Sun
We present a general boosting method extending functional gradient boosting to optimize complex loss functions that are encountered in many machine learning problems.
no code implementations • ICML 2017 • Wenpeng Zhang, Peilin Zhao, Wenwu Zhu, Steven C. H. Hoi, Tong Zhang
The conditional gradient algorithm has regained a surge of research interest in recent years due to its high efficiency in handling large-scale machine learning problems.
no code implementations • ICML 2018 • Qiang Sun, Kean Ming Tan, Han Liu, Tong Zhang
Our proposal is computationally tractable and produces an estimator that achieves the oracle rate of convergence.
no code implementations • ICLR 2019 • Yandong Li, Lijun Li, Liqiang Wang, Tong Zhang, Boqing Gong
In other words, there is a population of adversarial examples, instead of only one, for any input to a DNN.
no code implementations • 29 Dec 2018 • Haishan Ye, Zhichao Huang, Cong Fang, Chris Junchi Li, Tong Zhang
Zeroth-order optimization is an important research topic in machine learning.
no code implementations • 1 Feb 2019 • Cong Fang, Zhouchen Lin, Tong Zhang
In this paper, we give a sharp analysis for Stochastic Gradient Descent (SGD) and prove that SGD is able to efficiently escape from saddle points and find an $(\epsilon, O(\epsilon^{0. 5}))$-approximate second-order stationary point in $\tilde{O}(\epsilon^{-3. 5})$ stochastic gradient computations for generic nonconvex optimization problems, when the objective function satisfies gradient-Lipschitz, Hessian-Lipschitz, and dispersive noise assumptions.
no code implementations • 15 Feb 2019 • Zi-Yi Dou, Zhaopeng Tu, Xing Wang, Long-Yue Wang, Shuming Shi, Tong Zhang
With the promising progress of deep neural networks, layer aggregation has been used to fuse information across layers in various fields, such as computer vision and machine translation.
no code implementations • CVPR 2019 • Yinpeng Dong, Hang Su, Baoyuan Wu, Zhifeng Li, Wei Liu, Tong Zhang, Jun Zhu
In this paper, we evaluate the robustness of state-of-the-art face recognition models in the decision-based black-box attack setting, where the attackers have no access to the model parameters and gradients, but can only acquire hard-label predictions by sending queries to the target model.
no code implementations • 24 Apr 2019 • Tong Zhang, Pan Ji, Mehrtash Harandi, Wenbing Huang, Hongdong Li
We introduce the Neural Collaborative Subspace Clustering, a neural model that discovers clusters of data points drawn from a union of low-dimensional subspaces.
no code implementations • 15 May 2019 • Hanlin Tang, Xiangru Lian, Chen Yu, Tong Zhang, Ji Liu
For example, under the popular parameter server model for distributed learning, the worker nodes need to send the compressed local gradients to the parameter server, which performs the aggregation.
no code implementations • 28 May 2019 • Kean Ming Tan, Junwei Lu, Tong Zhang, Han Liu
To address this issue, neuroscientists have been measuring brain activity under natural viewing experiments in which the subjects are given continuous stimuli, such as watching a movie or listening to a story.
no code implementations • ICML 2018 • Li Shen, Peng Sun, Yitong Wang, Wei Liu, Tong Zhang
Specifically, we find that a large class of primal and primal-dual operator splitting algorithms are all special cases of VMOR-HPE.
no code implementations • 17 Jul 2019 • Hanlin Tang, Xiangru Lian, Shuang Qiu, Lei Yuan, Ce Zhang, Tong Zhang, Ji Liu
Since the \emph{decentralized} training has been witnessed to be superior to the traditional \emph{centralized} training in the communication restricted scenario, therefore a natural question to ask is "how to apply the error-compensated technology to the decentralized learning to further reduce the communication cost."
no code implementations • 28 Aug 2019 • Tong Zhang, Laurence H. Jackson, Alena Uus, James R. Clough, Lisa Story, Mary A. Rutherford, Joseph V. Hajnal, Maria Deprez
The results show that the proposed pipeline can accurately estimate the respiratory state and reconstruct 4D SR volumes with better or similar performance to the 3D SVR pipeline with less than 20\% sparsely selected slices.
no code implementations • 21 Oct 2019 • Ming-Han Yang, Andre Milzarek, Zaiwen Wen, Tong Zhang
In this paper, a novel stochastic extra-step quasi-Newton method is developed to solve a class of nonsmooth nonconvex composite optimization problems.
no code implementations • 25 Oct 2019 • Cong Fang, Hanze Dong, Tong Zhang
Recently, over-parameterized neural networks have been extensively analyzed in the literature.
no code implementations • 25 Oct 2019 • Haishan Ye, Tong Zhang
We show that the estimated covariance matrix of MiNES converges to the inverse of Hessian matrix of the objective function with a sublinear convergence rate.
no code implementations • 5 Nov 2019 • Jianchen Zhu, Tong Zhang, Shengjie Zhao, Carlos Hinojosa, Zengli Liu, Gonzalo R. Arce
This paper aims at developing a clustering approach with spectral images directly from CASSI compressive measurements.
no code implementations • 7 Nov 2019 • Tong Zhang, Fatih Porikli
The residual at a layer is computed by the difference between the aggregated reconstructions of the previous layers and the downsampled original image at that layer.
no code implementations • 18 Nov 2019 • Cong Fang, Yihong Gu, Weizhong Zhang, Tong Zhang
This new analysis is consistent with empirical observations that deep neural networks are capable of learning efficient feature representations.
no code implementations • 20 Nov 2019 • Daniel Chao Zhou, Zhongming Jin, Tong Zhang
As an adaptive, interpretable, robust, and accurate meta-algorithm for arbitrary differentiable loss functions, gradient tree boosting is one of the most popular machine learning techniques, though the computational expensiveness severely limits its usage.
no code implementations • 28 Nov 2019 • Zheyan Shen, Peng Cui, Tong Zhang, Kun Kuang
We consider the problem of learning linear prediction models with model misspecification bias.
no code implementations • 28 Nov 2019 • Xueya Zhang, Tong Zhang, Wenting Zhao, Zhen Cui, Jian Yang
Graph convolutional networks (GCNs) have shown the powerful ability in text structure representation and effectively facilitate the task of text classification.
no code implementations • 27 Dec 2019 • Haishan Ye, Shusen Wang, Zhihua Zhang, Tong Zhang
Fast matrix algorithms have become the fundamental tools of machine learning in big data era.
no code implementations • NeurIPS 2020 • Luo Luo, Haishan Ye, Zhichao Huang, Tong Zhang
We consider nonconvex-concave minimax optimization problems of the form $\min_{\bf x}\max_{\bf y\in{\mathcal Y}} f({\bf x},{\bf y})$, where $f$ is strongly-concave in $\bf y$ but possibly nonconvex in $\bf x$ and ${\mathcal Y}$ is a convex and compact set.
no code implementations • 15 Jan 2020 • Conghui Tan, Yuqiu Qian, Shiqian Ma, Tong Zhang
Dual averaging-type methods are widely used in industrial machine learning applications due to their ability to promoting solution structure (e. g., sparsity) efficiently.
no code implementations • ICLR 2020 • Chunyan Xu, Zhen Cui, Xiaobin Hong, Tong Zhang, Jian Yang, Wei Liu
In this work, we address semi-supervised classification of graph data, where the categories of those unlabeled nodes are inferred from labeled nodes as well as graph structures.
no code implementations • NeurIPS 2020 • Zixiang Chen, Yuan Cao, Quanquan Gu, Tong Zhang
In this paper, we provide a generalized neural tangent kernel analysis and show that noisy gradient descent with weight decay can still exhibit a "kernel-like" behavior.
no code implementations • 2 May 2020 • Haishan Ye, Luo Luo, Ziang Zhou, Tong Zhang
This paper considers the decentralized convex optimization problem, which has a wide range of applications in large-scale machine learning, sensor networks, and control theory.
no code implementations • ICML 2020 • Rie Johnson, Tong Zhang
This paper presents a framework of successive functional gradient optimization for training nonconvex models such as neural networks, where training is driven by mirror descent in a function space.
no code implementations • 3 Jul 2020 • Cong Fang, Jason D. Lee, Pengkun Yang, Tong Zhang
This new representation overcomes the degenerate situation where all the hidden units essentially have only one meaningful hidden unit in each middle layer, and further leads to a simpler representation of DNNs, for which the training objective can be reformulated as a convex optimization problem via suitable re-parameterization.
no code implementations • ECCV 2020 • Xin Chen, Yawen Duan, Zewei Chen, Hang Xu, Zihao Chen, Xiaodan Liang, Tong Zhang, Zhenguo Li
In spite of its remarkable progress, many algorithms are restricted to particular search spaces.
Ranked #13 on Neural Architecture Search on NAS-Bench-201, ImageNet-16-120 (Accuracy (Val) metric)
no code implementations • 26 Jul 2020 • Hongtao Yang, Tong Zhang, Wenbing Huang, Xuming He, Fatih Porikli
To be clear, in this paper, we refer unsupervised learning as learning without task-specific human annotations, pairs or any form of weak supervision.)
no code implementations • ECCV 2020 • Xueya Zhang, Tong Zhang, Xiaobin Hong, Zhen Cui, Jian Yang
Spectral graph filtering is introduced to encode graph signals, which are then embedded as probability distributions in a Wasserstein space, called graph Wasserstein metric learning.
no code implementations • 19 Aug 2020 • Yun Wang, Tong Zhang, Zhen Cui, Chunyan Xu, Jian Yang
For label diffusion of instance-awareness in graph convolution, rather than using the statistical label correlation alone, an image-dependent label correlation matrix (LCM), fusing both the statistical LCM and an individual one of each image instance, is constructed for graph inference on labels to inject adaptive information of label-awareness into the learned features of the model.
no code implementations • 24 Sep 2020 • Qianliang Wu, Tong Zhang, Zhen Cui, Jian Yang
In this paper, we aim to mine the cue of user preferences in resource-limited recommendation tasks, for which purpose we specifically build a large used car transaction dataset possessing resource-limitation characteristics.
no code implementations • 1 Jan 2021 • Xiao Zhou, Weizhong Zhang, Tong Zhang
An appealing feature of ProbMask is that the amounts of weight redundancy can be learned automatically via our constraint and thus we avoid the problem of tuning pruning rates individually for different layers in a network.
no code implementations • 1 Jan 2021 • Wenting Zhao, Yuan Fang, Zhen Cui, Tong Zhang, Jian Yang, Wei Liu
In this paper, we propose a simple yet effective graph deformer network (GDN) to fulfill anisotropic convolution filtering on graphs, analogous to the standard convolution operation on images.
no code implementations • 1 Jan 2021 • Qing Lian, LIN Yong, Tong Zhang
We consider the domain generalization problem, where the test domain differs from the training domain.
no code implementations • 3 Oct 2020 • Zhichao Huang, Yaowei Huang, Tong Zhang
We show that searching over the structured space can be approximated by a time-varying contextual bandits problem, where the attacker takes feature of the associated arm to make modifications of the input, and receives an immediate reward as the reduction of the loss function.
1 code implementation • NeurIPS 2020 • Yihong Gu, Weizhong Zhang, Cong Fang, Jason D. Lee, Tong Zhang
With the help of a new technique called {\it neural network grafting}, we demonstrate that even during the entire training process, feature distributions of differently initialized networks remain similar at each layer.
no code implementations • NeurIPS 2020 • Haishan Ye, Ziang Zhou, Luo Luo, Tong Zhang
In this paper, we propose a new method which establishes the optimal computational complexity and a near optimal communication complexity.
no code implementations • 15 Dec 2020 • Qi Chang, Zhennan Yan, Lohendran Baskaran, Hui Qu, Yikai Zhang, Tong Zhang, Shaoting Zhang, Dimitris N. Metaxas
As deep learning technologies advance, increasingly more data is necessary to generate general and robust models for various tasks.
no code implementations • 30 Dec 2020 • Haishan Ye, Wei Xiong, Tong Zhang
This paper considers the decentralized composite optimization problem.
no code implementations • 7 Dec 2020 • Wenqing Chen, Lulu Liu, Wentao Yang, Dong Chen, Zhengtai Liu, Yaobo Huang, Tong Zhang, Haijun Zhang, Zhonghao Liu, D. W. Shen
Utilizing angle-resolved photoemission spectroscopy and first-principles calculations, here, we demonstrate the existence of topological nodal-line states and drumheadlike surface states in centrosymmetric superconductor SnTaS2, which is a type-II superconductor with a critical transition temperature of about 3 K. The valence bands from Ta 5d orbitals and the conduction bands from Sn 5p orbitals cross each other, forming two nodal lines in the vicinity of the Fermi energy without the inclusion of spin-orbit coupling (SOC), protected by the spatial-inversion symmetry and time-reversal symmetry.
Superconductivity
no code implementations • 26 Jan 2021 • Min Yan, Guoshan Zhang, Tong Zhang, Yueming Zhang
In inference time, we design a brand-new grouping post-processing method that relates each part instance with one single human instance and groups them together to obtain the final human-level parsing result.
no code implementations • 8 Feb 2021 • Haishan Ye, Tong Zhang
This leads to a decentralized PCA algorithm called \texttt{DeEPCA}, which has a convergence rate similar to that of the centralized PCA, while achieving the best communication complexity among existing decentralized PCA algorithms.
no code implementations • 3 Jan 2021 • Tong Zhang, Yinfei Xu, Shuai Wang, Miaowen Wen, Rui Wang
This paper studies the problem of sum-secure degrees of freedom (SDoF) of the (M, M, N, N) multiple-input multiple-output (MIMO) interference channel with local output feedback, so as to build an information-theoretic foundation and provide practical transmission schemes for 6G-enabled vehicles-to-vehicles (V2V).
Information Theory Information Theory
no code implementations • 26 Nov 2008 • Daniel Hsu, Sham M. Kakade, Tong Zhang
Hidden Markov Models (HMMs) are one of the most fundamental and widely used statistical tools for modeling discrete time series.
1 code implementation • 1 Mar 2021 • Yang Yang, Jiancong Chen, Ruixuan Wang, Ting Ma, Lingwei Wang, Jie Chen, Wei-Shi Zheng, Tong Zhang
Despite tremendous efforts, it is very challenging to generate a robust model to assist in the accurate quantification assessment of COVID-19 on chest CT images.
no code implementations • 27 Feb 2021 • Wenrui Gan, Zhulin Liu, C. L. Philip Chen, Tong Zhang
In general, the main work of this paper include: (1) propose SiLa Learning, which improves the performance of common models without increasing test parameters; (2) compares SiLa with DML and proves that SiLa can improve the generalization of the model; (3) SiLa is applied to Dynamic Neural Networks, and proved that SiLa can be used for various types of network structures.
no code implementations • 10 Mar 2021 • Xuran Xu, Tong Zhang, Chunyan Xu, Zhen Cui, Jian Yang
We further extend graph convolution into tensor space and propose a tensor graph convolution network to extract more discriminating features from spatial-temporal graph data.
Ranked #1 on Traffic Prediction on SZ-Taxi
no code implementations • CVPR 2021 • Jie Hong, Pengfei Fang, Weihao Li, Tong Zhang, Christian Simon, Mehrtash Harandi, Lars Petersson
Few-shot learning aims to correctly recognize query samples from unseen classes given a limited number of support samples, often by relying on global embeddings of images.
no code implementations • CVPR 2022 • Qing Lian, Botao Ye, Ruijia Xu, Weilong Yao, Tong Zhang
In addition, we demonstrate that the augmentation methods are well suited for semi-supervised training and cross-dataset generalization.
no code implementations • 18 May 2021 • Tong Zhang, Yong liu, Peixiang Zhong, Chen Zhang, Hao Wang, Chunyan Miao
The chit-chat-based conversational recommendation systems (CRS) provide item recommendations to users through natural language interactions.
no code implementations • CVPR 2021 • Lewei Yao, Renjie Pi, Hang Xu, Wei zhang, Zhenguo Li, Tong Zhang
For student morphism, weight inheritance strategy is adopted, allowing the student to flexibly update its architecture while fully utilize the predecessor's weights, which considerably accelerates the search; To facilitate dynamic distillation, an elastic teacher pool is trained via integrated progressive shrinking strategy, from which teacher detectors can be sampled without additional cost in subsequent searches.
no code implementations • 29 May 2021 • Hanting Chen, Yunhe Wang, Chang Xu, Chao Xu, Chunjing Xu, Tong Zhang
The widely-used convolutions in deep neural networks are exactly cross-correlation to measure the similarity between input feature and convolution filters, which involves massive multiplications between float values.
no code implementations • 3 Jun 2021 • Luo Luo, Guangzeng Xie, Tong Zhang, Zhihua Zhang
This paper considers stochastic first-order algorithms for convex-concave minimax problems of the form $\min_{\bf x}\max_{\bf y}f(\bf x, \bf y)$, where $f$ can be presented by the average of $n$ individual components which are $L$-average smooth.
no code implementations • NAACL 2021 • Long Zhang, Tong Zhang, Haibo Zhang, Baosong Yang, Wei Ye, Shikun Zhang
Document-level neural machine translation (NMT) has proven to be of profound value for its effectiveness on capturing contextual information.
no code implementations • 20 Jul 2021 • Tong Zhang, Shuai Wang, Guoliang Li, Fan Liu, Guangxu Zhu, Rui Wang
Conventionally, the sensing and communication stages are executed sequentially, which results in excessive amount of dataset generation and uploading time.
no code implementations • ICCV 2021 • Lewei Yao, Renjie Pi, Hang Xu, Wei zhang, Zhenguo Li, Tong Zhang
In this paper, we investigate the knowledge distillation (KD) strategy for object detection and propose an effective framework applicable to both homogeneous and heterogeneous student-teacher pairs.
no code implementations • ACL 2021 • Tong Zhang, Long Zhang, Wei Ye, Bo Li, Jinan Sun, Xiaoyu Zhu, Wen Zhao, Shikun Zhang
This paper proposes a sophisticated neural architecture to incorporate bilingual dictionaries into Neural Machine Translation (NMT) models.
no code implementations • 20 Sep 2021 • Jieming Zhou, Tong Zhang, Pengfei Fang, Lars Petersson, Mehrtash Harandi
The core concept of GNNs is to find a representation by recursively aggregating the representations of a central node and those of its neighbors.
1 code implementation • 26 Sep 2021 • Jiancong Chen, Yingying Zhang, Jingyi Wang, Xiaoxue Zhou, Yihua He, Tong Zhang
In this paper, we present an anchor-free ellipse detection network, namely EllipseNet, which detects the cardiac and thoracic regions in ellipse and automatically calculates the CTR and cardiac axis for fetal cardiac biometrics in 4-chamber view.
no code implementations • 2 Oct 2021 • Tong Zhang
In this setting, we show that the standard Thompson Sampling is not aggressive enough in exploring new actions, leading to suboptimality in some pessimistic situations.
no code implementations • ICCV 2021 • Yun Wang, Tong Zhang, Xueya Zhang, Zhen Cui, Yuge Huang, Pengcheng Shen, Shaoxin Li, Jian Yang
Then, a Wasserstein coupled dictionary, containing multiple pairs of counterpart graph keys with each key corresponding to one modality, is constructed for further feature learning.
no code implementations • 5 Oct 2021 • Yoav Freund, Yi-An Ma, Tong Zhang
There has been a surge of works bridging MCMC sampling and optimization, with a specific focus on translating non-asymptotic convergence guarantees for optimization problems into the analysis of Langevin algorithms in MCMC sampling.
no code implementations • 29 Sep 2021 • Yanpeng Xie, Tong Zhang, Heng Zhang, Zhendong Qu
To make the framework model-agnostic, user Multi Interests Capsule Network is designed as an auxiliary task to jointly learn item-based item representations and interest-based item representations.
no code implementations • 29 Sep 2021 • Zhichao Huang, Chen Liu, Mathieu Salzmann, Sabine Süsstrunk, Tong Zhang
Although adversarial training and its variants currently constitute the most effective way to achieve robustness against adversarial attacks, their poor generalization limits their performance on the test samples.
no code implementations • 27 Sep 2018 • Miaofeng Liu, Yan Song, Hongbin Zou, Tong Zhang
Following the TDS methodology, in this paper, we propose a general data selection framework with representation learning and distribution matching simultaneously for domain adaptation on neural models.
no code implementations • 25 Sep 2019 • Han Shi, Renjie Pi, Hang Xu, Zhenguo Li, James T. Kwok, Tong Zhang
Inspired by the nature of the graph structure of a neural network, we propose BOGCN-NAS, a NAS algorithm using Bayesian Optimization with Graph Convolutional Network (GCN) predictor.
no code implementations • 14 Dec 2021 • Chen Liu, Zhichao Huang, Mathieu Salzmann, Tong Zhang, Sabine Süsstrunk
This lets us show that the decay in generalization performance of adversarial training is a result of the model's attempt to fit hard adversarial instances.
no code implementations • 29 Dec 2021 • Tong Zhang, Wei Ye, Baosong Yang, Long Zhang, Xingzhang Ren, Dayiheng Liu, Jinan Sun, Shikun Zhang, Haibo Zhang, Wen Zhao
Inspired by the observation that low-frequency words form a more compact embedding space, we tackle this challenge from a representation learning perspective.
no code implementations • 14 Jan 2022 • Mengyue Zha, Kani Chen, Tong Zhang
We enhance the accuracy and generalization of univariate time series point prediction by an explainable ensemble on the fly.
no code implementations • 15 Jan 2022 • Tong Zhang, Haohan Weng, Ke Yi, C. L. Philip Chen
Convolutional Neural Networks (CNNs) have exhibited their great power in a variety of vision tasks.
no code implementations • 15 Jan 2022 • Jibao Qiu, C. L. Philip Chen, Tong Zhang
In this paper, we present a simple multi-task framework for SMER, which incorporates the emotion recognition task with other emotion-related auxiliary tasks derived from the intrinsic structure of the music.
no code implementations • 11 Feb 2022 • Claudio Gentile, Zhilei Wang, Tong Zhang
We consider a batch active learning scenario where the learner adaptively issues batches of points to a labeling oracle.
no code implementations • 11 Feb 2022 • Alekh Agarwal, Tong Zhang
We instead propose an alternative method called Minimax Regret Optimization (MRO), and show that under suitable conditions this method achieves uniformly low regret across all test distributions.
no code implementations • 15 Feb 2022 • Han Zhong, Wei Xiong, Jiyuan Tan, LiWei Wang, Tong Zhang, Zhaoran Wang, Zhuoran Yang
When the dataset does not have uniform coverage over all policy pairs, finding an approximate NE involves challenges in three aspects: (i) distributional shift between the behavior policy and the optimal policy, (ii) function approximation to handle large state space, and (iii) minimax optimization for equilibrium solving.
no code implementations • 15 Mar 2022 • Alekh Agarwal, Tong Zhang
Provably sample-efficient Reinforcement Learning (RL) with rich observations and function approximation has witnessed tremendous recent progress, particularly when the underlying function approximators are linear.
no code implementations • CVPR 2022 • Tong Zhang, Congpei Qiu, Wei Ke, Sabine Süsstrunk, Mathieu Salzmann
In essence, this strategy ignores the fact that two crops may truly contain different image information, e. g., background and small objects, and thus tends to restrain the diversity of the learned representations.
no code implementations • 10 Apr 2022 • Hui Deng, Tong Zhang, Yuchao Dai, Jiawei Shi, Yiran Zhong, Hongdong Li
In this paper, we propose to model deep NRSfM from a sequence-to-sequence translation perspective, where the input 2D frame sequence is taken as a whole to reconstruct the deforming 3D non-rigid shape sequence.
no code implementations • NLP4ConvAI (ACL) 2022 • Tong Zhang, Yong liu, Boyang Li, Peixiang Zhong, Chen Zhang, Hao Wang, Chunyan Miao
Conversational Recommendation Systems recommend items through language based interactions with users. In order to generate naturalistic conversations and effectively utilize knowledge graphs (KGs) containing background information, we propose a novel Bag-of-Entities loss, which encourages the generated utterances to mention concepts related to the item being recommended, such as the genre or director of a movie.
no code implementations • 13 May 2022 • Jiafan He, Dongruo Zhou, Tong Zhang, Quanquan Gu
We show that for both known $C$ and unknown $C$ cases, our algorithm with proper choice of hyperparameter achieves a regret that nearly matches the lower bounds.
no code implementations • CVPR 2022 • Deblina Bhattacharjee, Tong Zhang, Sabine Süsstrunk, Mathieu Salzmann
At the heart of our approach is a shared attention mechanism modeling the dependencies across the tasks.
no code implementations • 31 May 2022 • Wei Xiong, Han Zhong, Chengshuai Shi, Cong Shen, LiWei Wang, Tong Zhang
We also extend our techniques to the two-player zero-sum Markov games (MGs), and establish a new performance lower bound for MGs, which tightens the existing result, and verifies the nearly minimax optimality of the proposed algorithm.
no code implementations • CVPR 2022 • Yong Lin, Hanze Dong, Hao Wang, Tong Zhang
Generalization under distributional shift is an open challenge for machine learning.
no code implementations • 3 Jun 2022 • Yi-An Ma, Teodor Vanislavov Marinov, Tong Zhang
This paper considers the generalization performance of differentially private convex learning.
no code implementations • 7 Jun 2022 • Yifei Sun, Jie Li, Tong Zhang, Rui Wang, Xiaohui Peng, Tony Xiao Han, Haisheng Tan
At the end, we show that the reconstructed room layout can be utilized to locate a mobile device according to its AoA spectrum, even with single access point.
no code implementations • 9 Jun 2022 • Hao liu, Minshuo Chen, Siawpeng Er, Wenjing Liao, Tong Zhang, Tuo Zhao
Overparameterized neural networks enjoy great representation power on complex data, and more importantly yield sufficiently smooth output, which is crucial to their generalization and robustness.
no code implementations • 9 Jun 2022 • Jianyu Wang, Rudrajit Das, Gauri Joshi, Satyen Kale, Zheng Xu, Tong Zhang
Motivated by this observation, we propose a new quantity, average drift at optimum, to measure the effects of data heterogeneity, and explicitly use it to present a new theoretical analysis of FedAvg.
no code implementations • 15 Jun 2022 • Alekh Agarwal, Tong Zhang
We propose a general framework to design posterior sampling methods for model-based RL.
no code implementations • 21 Jun 2022 • Rudrajit Das, Satyen Kale, Zheng Xu, Tong Zhang, Sujay Sanghavi
Most prior results on differentially private stochastic gradient descent (DP-SGD) are derived under the simplistic assumption of uniform Lipschitzness, i. e., the per-sample gradients are uniformly bounded.
no code implementations • IEEE Transactions on Intelligent Transportation Systems 2022 • Junchen Jin, Member, IEEE, Dingding Rong, Tong Zhang, Qingyuan Ji, Haifeng Guo, Yisheng Lv, Xiaoliang Ma, and Fei-Yue Wang
This paper proposes a short-term traffic speed prediction approach, called PL-WGAN, for urban road networks, which is considered an important part of a novel parallel learning framework for traffic control and operation.
no code implementations • 11 Aug 2022 • Jia-Xin Zhuang, Xiansong Huang, Yang Yang, Jiancong Chen, Yue Yu, Wei Gao, Ge Li, Jie Chen, Tong Zhang
In this paper, we present OpenMedIA, an open-source toolbox library containing a rich set of deep learning methods for medical image analysis under heterogeneous Artificial Intelligence (AI) computing platforms.
no code implementations • NeurIPS 2021 • Christoph Dann, Mehryar Mohri, Tong Zhang, Julian Zimmert
Thompson Sampling is one of the most effective methods for contextual bandits and has been generalized to posterior sampling for certain MDP settings.
no code implementations • COLING 2022 • Zile Qiao, Wei Ye, Tong Zhang, Tong Mo, Weiping Li, Shikun Zhang
Answering natural language questions on knowledge graphs (KGQA) remains a great challenge in terms of understanding complex questions via multi-hop reasoning.
no code implementations • 14 Sep 2022 • Xinwei Shen, Kani Chen, Tong Zhang
We show that for parametric generative models that are correctly specified, all $f$-divergence GANs with the same discriminator classes are asymptotically equivalent under suitable regularity conditions.
no code implementations • 27 Sep 2022 • Tong Zhang, Christopher Williams, Reza Ahmadian, Meysam Qadrdan
It was demonstrated that by exploiting the flexibility offered by the tidal lagoon, it can achieve a higher revenue in the day-ahead market, although their total electricity generation is reduced.
no code implementations • 4 Oct 2022 • Wei Xiong, Han Zhong, Chengshuai Shi, Cong Shen, Tong Zhang
Existing studies on provably efficient algorithms for Markov games (MGs) almost exclusively build on the "optimism in the face of uncertainty" (OFU) principle.
1 code implementation • COLING 2022 • Ying Su, Hongming Zhang, Yangqiu Song, Tong Zhang
As a key natural language processing (NLP) task, word sense disambiguation (WSD) evaluates how well NLP models can understand the lexical semantics of words under specific contexts.
no code implementations • 16 Oct 2022 • Baijun Ji, Tong Zhang, Yicheng Zou, Bojie Hu, Si Shen
Multimodal machine translation (MMT) aims to improve translation quality by equipping the source sentence with its corresponding image.
no code implementations • 18 Oct 2022 • Xinrao Li, Tong Zhang, Shuai Wang, Guangxu Zhu, Rui Wang, Tsung-Hui Chang
However, wireless channels between the edge server and the autonomous vehicles are time-varying due to the high-mobility of vehicles.
no code implementations • 3 Nov 2022 • Han Zhong, Wei Xiong, Sirui Zheng, LiWei Wang, Zhaoran Wang, Zhuoran Yang, Tong Zhang
The proposed algorithm modifies the standard posterior sampling algorithm in two aspects: (i) we use an optimistic prior distribution that biases towards hypotheses with higher values and (ii) a loglikelihood function is set to be the empirical loss evaluated on the historical data, where the choice of loss function supports both model-free and model-based learning.
no code implementations • 11 Nov 2022 • Kilean Hwang, Tomofumi Maruta, Alexander Plastun, Kei Fukushima, Tong Zhang, Qiang Zhao, Peter Ostroumov, Yue Hao
Bayesian optimization~(BO) is often used for accelerator tuning due to its high sample efficiency.
no code implementations • CVPR 2023 • Ehsan Pajouheshgar, Yitao Xu, Tong Zhang, Sabine Süsstrunk
Current Dynamic Texture Synthesis (DyTS) models can synthesize realistic videos.
no code implementations • 25 Nov 2022 • Hanze Dong, Xi Wang, Yong Lin, Tong Zhang
With the popularity of Stein variational gradient descent (SVGD), the focus of particle-based VI algorithms has been on the properties of functions in Reproducing Kernel Hilbert Space (RKHS) to approximate the gradient flow.
no code implementations • 20 Nov 2022 • Zhongyu Fang, Aoyun He, Qihui Yu, Baopeng Gao, Weiping Ding, Tong Zhang, Lei Ma
In this paper, we developed a large multimodal emotion dataset, named "HED" dataset, to facilitate the emotion recognition task, and accordingly propose a multimodal emotion recognition method.
no code implementations • 29 Nov 2022 • Tong Zhang, Ying Tan, Xiang Chen, Zike Lei
The key design idea for this observer is to estimate the visible set and identify the mis-identified features from the measurements.
no code implementations • 12 Dec 2022 • Chenlu Ye, Wei Xiong, Quanquan Gu, Tong Zhang
In this paper, we consider the contextual bandit with general function approximation and propose a computationally efficient algorithm to achieve a regret of $\tilde{O}(\sqrt{T}+\zeta)$.
no code implementations • 12 Dec 2022 • Alekh Agarwal, Yujia Jin, Tong Zhang
We study time-inhomogeneous episodic reinforcement learning (RL) under general function approximation and sparse rewards.