no code implementations • NeurIPS 2007 • Zhaohui Zheng, Hongyuan Zha, Tong Zhang, Olivier Chapelle, Keke Chen, Gordon Sun
We present a general boosting method extending functional gradient boosting to optimize complex loss functions that are encountered in many machine learning problems.
no code implementations • NeurIPS 2007 • John Langford, Tong Zhang
We present Epoch-Greedy, an algorithm for multi-armed bandits with observable side information.
no code implementations • 26 Nov 2008 • Daniel Hsu, Sham M. Kakade, Tong Zhang
Hidden Markov Models (HMMs) are one of the most fundamental and widely used statistical tools for modeling discrete time series.
no code implementations • NeurIPS 2008 • John Langford, Lihong Li, Tong Zhang
We propose a general method called truncated gradient to induce sparsity in the weights of online-learning algorithms with convex loss.
no code implementations • NeurIPS 2008 • Tong Zhang
We study learning formulations with non-convex regularizaton that are natural for sparse linear models.
no code implementations • NeurIPS 2008 • Tong Zhang
Consider linear prediction models where the target function is a sparse linear combination of a set of basis functions.
no code implementations • NeurIPS 2009 • Kai Yu, Tong Zhang, Yihong Gong
This paper introduces a new method for semi-supervised learning on high dimensional nonlinear manifolds, which includes a phase of unsupervised basis learning and a phase of supervised function learning.
no code implementations • NeurIPS 2010 • Alina Beygelzimer, Daniel J. Hsu, John Langford, Tong Zhang
We present and analyze an agnostic active learning algorithm that works without keeping a version space.
no code implementations • NeurIPS 2010 • Yuanqing Lin, Tong Zhang, Shenghuo Zhu, Kai Yu
This paper proposes a principled extension of the traditional single-layer flat sparse coding scheme, where a two-layer coding scheme is derived based on theoretical analysis of nonlinear functional approximation that extends recent results for local coordinate coding.
no code implementations • 13 Jun 2011 • Daniel Hsu, Sham M. Kakade, Tong Zhang
The analysis also reveals the effect of errors in the estimated covariance structure, as well as the effect of modeling errors, neither of which effects are present in the fixed design setting.
1 code implementation • 5 Sep 2011 • Rie Johnson, Tong Zhang
We consider the problem of learning a forest of nonlinear decision rules with general loss functions.
no code implementations • NeurIPS 2011 • Animashree Anandkumar, Kamalika Chaudhuri, Daniel J. Hsu, Sham M. Kakade, Le Song, Tong Zhang
The setting is one where we only have samples from certain observed variables in the tree, and our goal is to estimate the tree structure (i. e., the graph of how the underlying hidden variables are connected to each other and to the observed variables).
no code implementations • NeurIPS 2011 • Zhen Li, Huazhong Ning, Liangliang Cao, Tong Zhang, Yihong Gong, Thomas S. Huang
Traditional approaches relied on algorithmic constructions that are often data independent (such as Locality Sensitive Hashing) or weakly dependent (such as kd-trees, k-means trees).
no code implementations • NeurIPS 2011 • Dong Dai, Tong Zhang
The purpose of this paper is to present a new greedy model averaging procedure that improves EWMA.
no code implementations • NeurIPS 2012 • Quanquan Gu, Tong Zhang, Jiawei Han, Chris H. Ding
In particular, we derive a deterministic generalization error bound for LapRLS trained on subsampled data, and propose to select a subset of data points to label by minimizing this upper bound.
no code implementations • 13 Dec 2012 • Sivan Sabato, Shai Shalev-Shwartz, Nathan Srebro, Daniel Hsu, Tong Zhang
We consider the problem of learning a non-negative linear classifier with a $1$-norm of at most $k$, and a fixed threshold, under the hinge-loss.
no code implementations • NeurIPS 2013 • Shai Shalev-Shwartz, Tong Zhang
Stochastic dual coordinate ascent (SDCA) is an effective technique for solving regularized loss minimization problems in machine learning.
no code implementations • 20 Jun 2013 • Zhaoran Wang, Han Liu, Tong Zhang
In particular, our analysis improves upon existing results by providing a more refined sample complexity bound as well as an exact support recovery result for the final estimator.
no code implementations • 10 Sep 2013 • Shai Shalev-Shwartz, Tong Zhang
We introduce a proximal version of the stochastic dual coordinate ascent method and show how to accelerate the method using an inner-outer iteration procedure.
no code implementations • 26 Sep 2013 • Krishnakumar Balasubramanian, Kai Yu, Tong Zhang
The traditional convex formulation employs the group Lasso relaxation to achieve joint sparsity across tasks.
no code implementations • 3 Oct 2013 • Ping Li, Cun-Hui Zhang, Tong Zhang
In particular, when p->0 the required number of measurements is essentially M=K\log N, where K is the number of nonzero coordinates of the signal.
no code implementations • 12 Nov 2013 • Dong Dai, Philippe Rigollet, Lucy Xia, Tong Zhang
While results indicate that the same aggregation scheme may not satisfy sharp oracle inequalities with high probability, we prove that a weaker notion of oracle inequality for EW that holds with high probability.
no code implementations • 21 Nov 2013 • Xiao-Tong Yuan, Ping Li, Tong Zhang
We investigate a generic problem of learning pairwise exponential family graphical models with pairwise sufficient statistics defined by a global mapping function, e. g., Mercer kernels.
no code implementations • 22 Nov 2013 • Xiao-Tong Yuan, Ping Li, Tong Zhang
Numerical evidences show that our method is superior to the state-of-the-art greedy selection methods in sparse logistic regression and sparse precision matrix estimation tasks.
no code implementations • NeurIPS 2013 • Rie Johnson, Tong Zhang
Stochastic gradient descent is popular for large scale optimization but has slow convergence asymptotically due to the inherent variance.
no code implementations • 16 Dec 2013 • Peilin Zhao, Jinwei Yang, Tong Zhang, Ping Li
The Alternating Direction Method of Multipliers (ADMM) has been studied for years.
1 code implementation • 30 Dec 2013 • Ohad Shamir, Nathan Srebro, Tong Zhang
We present a novel Newton-type method for distributed optimization, which is particularly well suited for stochastic optimization and learning problems.
no code implementations • 31 Dec 2013 • Ping Li, Cun-Hui Zhang, Tong Zhang
In this paper, we adopt very sparse Compressed Counting for nonnegative signal recovery.
no code implementations • 13 Jan 2014 • Peilin Zhao, Tong Zhang
Uniform sampling of training data has been commonly used in traditional stochastic optimization algorithms such as Proximal Stochastic Gradient Descent (prox-SGD) and Proximal Stochastic Dual Coordinate Ascent (prox-SDCA).
no code implementations • 19 Mar 2014 • Lin Xiao, Tong Zhang
We consider the problem of minimizing the sum of two convex functions: one is the average of a large number of smooth component functions, and the other is a general convex function that admits a simple proximal mapping.
no code implementations • 13 May 2014 • Peilin Zhao, Tong Zhang
Stochastic Gradient Descent (SGD) is a popular optimization method which has been applied to many important machine learning tasks such as Support Vector Machines and Deep Neural Networks.
no code implementations • 21 Nov 2014 • Zheng Qu, Peter Richtárik, Tong Zhang
The distributed variant of Quartz is the first distributed SDCA-like method with an analysis for non-separable data.
4 code implementations • HLT 2015 • Rie Johnson, Tong Zhang
Convolutional neural network (CNN) is a neural network that can make use of the internal structure of data such as the 2D structure of image data.
Ranked #29 on Sentiment Analysis on IMDb
no code implementations • 23 Dec 2014 • Tuo Zhao, Han Liu, Tong Zhang
This is the first result on the computational and statistical guarantees of the pathwise coordinate optimization framework in high dimensions.
no code implementations • 26 Dec 2014 • Shusen Wang, Tong Zhang, Zhihua Zhang
Low-rank matrix completion is an important problem with extensive real-world applications.
no code implementations • 29 Mar 2015 • Shusen Wang, Zhihua Zhang, Tong Zhang
The Nystr\"om method is a special instance of our fast model and is approximation to the prototype model.
no code implementations • NeurIPS 2015 • Rie Johnson, Tong Zhang
This paper presents a new semi-supervised framework with convolutional neural networks (CNNs) for text categorization.
Ranked #1000000000 on Text Classification on IMDb
no code implementations • 14 Nov 2015 • Zhuoran Yang, Zhaoran Wang, Han Liu, Yonina C. Eldar, Tong Zhang
To recover $\beta^*$, we propose an $\ell_1$-regularized least-squares estimator.
no code implementations • NeurIPS 2015 • Zheng Qu, Peter Richtarik, Tong Zhang
We study the problem of minimizing the average of a large number of smooth convex functions penalized with a strongly convex regularizer.
no code implementations • NeurIPS 2015 • Daniel Vainsencher, Han Liu, Tong Zhang
Abstract We propose a family of non-uniform sampling strategies to provably speed up a class of stochastic optimization algorithms with linear convergence including Stochastic Variance Reduced Gradient (SVRG) and Stochastic Dual Coordinate Ascent (SDCA).
no code implementations • 7 Feb 2016 • Rie Johnson, Tong Zhang
The best results were obtained by combining region embeddings in the form of LSTM and convolution layers trained on unlabeled data.
Ranked #1 on Text Classification on RCV1
no code implementations • 16 Mar 2016 • Chris Junchi Li, Mengdi Wang, Han Liu, Tong Zhang
We prove for the first time a nearly optimal finite-sample error bound for the online PCA algorithm.
no code implementations • 13 Apr 2016 • Shun Zheng, Jialei Wang, Fen Xia, Wei Xu, Tong Zhang
In modern large-scale machine learning applications, the training data are often partitioned and stored on multiple machines.
no code implementations • 27 Apr 2016 • Lei Han, Kean Ming Tan, Ting Yang, Tong Zhang
A major challenge for building statistical models in the big data era is that the available data volume far exceeds the computational capability.
no code implementations • 29 Apr 2016 • Kean Ming Tan, Zhaoran Wang, Han Liu, Tong Zhang
Sparse generalized eigenvalue problem (GEP) plays a pivotal role in a large family of high-dimensional statistical models, including sparse Fisher's discriminant analysis, canonical correlation analysis, and sufficient dimension reduction.
no code implementations • ICML 2017 • Jialei Wang, Mladen Kolar, Nathan Srebro, Tong Zhang
We propose a novel, efficient approach for distributed sparse learning in high-dimensions, where observations are randomly partitioned across machines.
no code implementations • 31 Aug 2016 • Rie Johnson, Tong Zhang
This paper reports the performances of shallow word-level convolutional neural networks (CNN), our earlier work (2015), on the eight datasets with relatively large training data that were used for testing the very deep character-level CNN in Conneau et al. (2016).
no code implementations • NeurIPS 2016 • Xiaotong Yuan, Ping Li, Tong Zhang
In this paper, we bridge this gap by showing, for the first time, that exact recovery of the global sparse minimizer is possible for HTP-style methods under restricted strong condition number bounding conditions.
no code implementations • NeurIPS 2016 • Xiaotong Yuan, Ping Li, Tong Zhang, Qingshan Liu, Guangcan Liu
We investigate a subclass of exponential family graphical models of which the sufficient statistics are defined by arbitrary additive forms.
1 code implementation • 8 Feb 2017 • Lingke Zeng, Xiangmin Xu, Bolun Cai, Suo Qiu, Tong Zhang
Crowd counting on static images is a challenging problem due to scale variations.
no code implementations • 12 May 2017 • Tong Zhang, Wenming Zheng, Zhen Cui, Yuan Zong, Yang Li
Then a bi-directional temporal RNN layer is further used to learn discriminative temporal dependencies from the sequences concatenating spatial features of each time slice produced from the spatial RNN layer.
no code implementations • 15 May 2017 • Luo Luo, Cheng Chen, Zhihua Zhang, Wu-Jun Li, Tong Zhang
We also apply RFD to online learning and propose an effective hyperparameter-free online Newton algorithm.
no code implementations • 30 May 2017 • Tong Zhang, Wenming Zheng, Zhen Cui, Chaolong Li
Symmetric positive definite (SPD) matrices (e. g., covariances, graph Laplacians, etc.)
no code implementations • ICML 2018 • Wenhan Luo, Peng Sun, Fangwei Zhong, Wei Liu, Tong Zhang, Yizhou Wang
We study active object tracking, where a tracker takes as input the visual observation (i. e., frame sequence) and produces the camera control signal (e. g., move forward, turn left, etc.).
no code implementations • 4 Jun 2017 • Qiang Sun, Kean Ming Tan, Han Liu, Tong Zhang
Our proposal is computationally tractable and produces an estimator that achieves the oracle rate of convergence.
no code implementations • 19 Jun 2017 • Xingguo Li, Lin F. Yang, Jason Ge, Jarvis Haupt, Tong Zhang, Tuo Zhao
We propose a DC proximal Newton algorithm for solving nonconvex regularized sparse learning problems in high dimensions.
no code implementations • 21 Jun 2017 • Jialei Wang, Tong Zhang
We present novel minibatch stochastic optimization methods for empirical risk minimization problems, the methods efficiently leverage variance reduced first-order and sub-sampled higher-order information to accelerate the convergence speed.
1 code implementation • ACL 2017 • Rie Johnson, Tong Zhang
This paper proposes a low-complexity word-level deep convolutional neural network (CNN) architecture for text categorization that can efficiently represent long-range associations in text.
Ranked #2 on Sentiment Analysis on Amazon Review Full
no code implementations • ICML 2017 • Wenpeng Zhang, Peilin Zhao, Wenwu Zhu, Steven C. H. Hoi, Tong Zhang
The conditional gradient algorithm has regained a surge of research interest in recent years due to its high efficiency in handling large-scale machine learning problems.
3 code implementations • NeurIPS 2017 • Pan Ji, Tong Zhang, Hongdong Li, Mathieu Salzmann, Ian Reid
We present a novel deep neural network architecture for unsupervised subspace clustering.
Ranked #3 on Image Clustering on Extended Yale-B
no code implementations • NeurIPS 2018 • Jianqiao Wangni, Jialei Wang, Ji Liu, Tong Zhang
Modern large scale machine learning applications require stochastic optimization algorithms to be implemented on distributed computational architectures.
no code implementations • ICML 2018 • Lei Han, Yiheng Huang, Tong Zhang
This paper proposes a method for multi-class classification problems, where the number of classes K is large.
1 code implementation • TACL 2018 • Zhaopeng Tu, Yang Liu, Shuming Shi, Tong Zhang
Existing neural machine translation (NMT) models generally translate sentences in isolation, missing the opportunity to take advantage of document-level information.
no code implementations • NeurIPS 2017 • Wenbing Huang, Mehrtash Harandi, Tong Zhang, Lijie Fan, Fuchun Sun, Junzhou Huang
Linear Dynamical Systems (LDSs) are fundamental tools for modeling spatio-temporal data in various disciplines.
no code implementations • NeurIPS 2017 • Xingguo Li, Lin Yang, Jason Ge, Jarvis Haupt, Tong Zhang, Tuo Zhao
We propose a DC proximal Newton algorithm for solving nonconvex regularized sparse learning problems in high dimensions.
1 code implementation • ICLR 2018 • Jiechao Xiong, Qing Wang, Zhuoran Yang, Peng Sun, Yang Zheng, Lei Han, Haobo Fu, Xiangru Lian, Carson Eisenach, Haichuan Yang, Emmanuel Ekwedike, Bei Peng, Haoyue Gao, Tong Zhang, Ji Liu, Han Liu
Most existing deep reinforcement learning (DRL) frameworks consider action spaces that are either discrete or continuous space.
1 code implementation • 10 Jan 2018 • Long-Yue Wang, Zhaopeng Tu, Shuming Shi, Tong Zhang, Yvette Graham, Qun Liu
Next, the annotated source sentence is reconstructed from hidden representations in the NMT model.
no code implementations • ICML 2018 • Rie Johnson, Tong Zhang
This paper first presents a theory for generative adversarial methods that does not rely on the traditional minimax formulation.
5 code implementations • ICML 2018 • Kaiqing Zhang, Zhuoran Yang, Han Liu, Tong Zhang, Tamer Başar
To this end, we propose two decentralized actor-critic algorithms with function approximation, which are applicable to large-scale MARL problems where both the number of states and the number of agents are massively large.
Multi-agent Reinforcement Learning reinforcement-learning +1
no code implementations • ECCV 2018 • Xinyu Gong, HaoZhi Huang, Lin Ma, Fumin Shen, Wei Liu, Tong Zhang
While each view of the stereoscopic pair is processed in an individual path, a novel feature aggregation strategy is proposed to effectively share information between the two paths.
no code implementations • NeurIPS 2018 • Hanlin Tang, Shaoduo Gan, Ce Zhang, Tong Zhang, Ji Liu
In this paper, We explore a natural question: {\em can the combination of both techniques lead to a system that is robust to both bandwidth and latency?}
no code implementations • 27 Mar 2018 • Tong Zhang, Wenming Zheng, Zhen Cui, Yang Li
For cross graph convolution, a parameterized Kronecker sum operation is proposed to generate a conjunctive adjacency matrix characterizing the relationship between every pair of nodes across two subgraphs.
no code implementations • CVPR 2018 • Jing Zhang, Tong Zhang, Yuchao Dai, Mehrtash Harandi, Richard Hartley
Such supervision, while labor-intensive and not always possible, tends to hinder the generalization ability of the learned models.
no code implementations • 4 Apr 2018 • Xinpeng Chen, Jingyuan Chen, Lin Ma, Jian Yao, Wei Liu, Jiebo Luo, Tong Zhang
First, we demonstrate that video attractiveness and different engagements present different relationships.
no code implementations • 16 Apr 2018 • Jiatao Jiang, Chunyan Xu, Zhen Cui, Tong Zhang, Wenming Zheng, Jian Yang
As an analogy to a standard convolution kernel on image, Gaussian models implicitly coordinate those unordered vertices/nodes and edges in a local receptive field after projecting to the gradient space of Gaussian parameters.
1 code implementation • EMNLP 2018 • Yi Liao, Lidong Bing, Piji Li, Shuming Shi, Wai Lam, Tong Zhang
For example, an input sequence could be a word sequence, such as review sentence and advertisement text.
no code implementations • ICML 2018 • Li Shen, Peng Sun, Yitong Wang, Wei Liu, Tong Zhang
Specifically, we find that a large class of primal and primal-dual operator splitting algorithms are all special cases of VMOR-HPE.
no code implementations • ICML 2018 • Weizhong Zhang, Bin Hong, Lin Ma, Wei Liu, Tong Zhang
Relying on this study, we subsequently propose a novel safe screening method to quickly identify the elements guaranteed to be included (we refer to them as active) or excluded (inactive) in the final optimal solution of SFM during the optimization process.
no code implementations • ICML 2018 • Jiaxiang Wu, Weidong Huang, Junzhou Huang, Tong Zhang
Large-scale distributed optimization is of great importance in various applications.
no code implementations • ICML 2018 • Qiang Sun, Kean Ming Tan, Han Liu, Tong Zhang
Our proposal is computationally tractable and produces an estimator that achieves the oracle rate of convergence.
no code implementations • NeurIPS 2018 • Cong Fang, Chris Junchi Li, Zhouchen Lin, Tong Zhang
For stochastic first-order method, combining SPIDER with normalized gradient descent, we propose two new algorithms, namely SPIDER-SFO and SPIDER-SFO\textsuperscript{+}, that solve non-convex stochastic optimization problems using stochastic gradients only.
no code implementations • 7 Jul 2018 • Wenting Zhao, Chunyan Xu, Zhen Cui, Tong Zhang, Jiatao Jiang, Zhen-Yu Zhang, Jian Yang
In this paper, we aim to give a comprehensive analysis of when work matters by transforming different classical network structures to graph CNN, particularly in the basic graph recognition problem.
Ranked #3 on Graph Classification on IMDb-B
no code implementations • ECCV 2018 • Minjun Li, Hao-Zhi Huang, Lin Ma, Wei Liu, Tong Zhang, Yu-Gang Jiang
Recent studies on unsupervised image-to-image translation have made a remarkable progress by training a pair of generative adversarial networks with a cycle-consistent loss.
no code implementations • ECCV 2018 • Wenhao Jiang, Lin Ma, Yu-Gang Jiang, Wei Liu, Tong Zhang
In this paper, in order to exploit the complementary information from multiple encoders, we propose a novel Recurrent Fusion Network (RFNet) for tackling image captioning.
1 code implementation • ECCV 2018 • Yang Feng, Lin Ma, Wei Liu, Tong Zhang, Jiebo Luo
We first exploit and reorganize the videos in ActivityNet to form a new dataset for video re-localization research, which consists of about 10, 000 videos of diverse visual appearances associated with localized boundary information.
no code implementations • 10 Aug 2018 • Wenhan Luo, Peng Sun, Fangwei Zhong, Wei Liu, Tong Zhang, Yizhou Wang
We further propose an environment augmentation technique and a customized reward function, which are crucial for successful training.
no code implementations • NeurIPS 2017 • Chris Junchi Li, Mengdi Wang, Han Liu, Tong Zhang
In this paper, we propose to adopt the diffusion approximation tools to study the dynamics of Oja's iteration which is an online stochastic gradient descent method for the principal component analysis.
2 code implementations • NeurIPS 2018 • Wenbing Huang, Tong Zhang, Yu Rong, Junzhou Huang
Graph Convolutional Networks (GCNs) have become a crucial tool on learning representations of graph vertices.
Ranked #2 on Node Classification on Pubmed Full-supervised
no code implementations • 17 Sep 2018 • Kean Ming Tan, Zhaoran Wang, Tong Zhang, Han Liu, R. Dennis Cook
Sliced inverse regression is a popular tool for sufficient dimension reduction, which replaces covariates with a minimal set of their linear combinations without loss of information on the conditional distribution of the response given the covariates.
3 code implementations • 19 Sep 2018 • Peng Sun, Xinghai Sun, Lei Han, Jiechao Xiong, Qing Wang, Bo Li, Yang Zheng, Ji Liu, Yongsheng Liu, Han Liu, Tong Zhang
Both TStarBot1 and TStarBot2 are able to defeat the built-in AI agents from level 1 to level 10 in a full game (1v1 Zerg-vs-Zerg game on the AbyssalReef map), noting that level 8, level 9, and level 10 are cheating agents with unfair advantages such as full vision on the whole map and resource harvest boosting.
no code implementations • 25 Sep 2018 • Chaobing Song, Ji Liu, Han Liu, Yong Jiang, Tong Zhang
Regularized online learning is widely used in machine learning applications.
no code implementations • 27 Sep 2018 • Miaofeng Liu, Yan Song, Hongbin Zou, Tong Zhang
Following the TDS methodology, in this paper, we propose a general data selection framework with representation learning and distribution matching simultaneously for domain adaptation on neural models.
1 code implementation • EMNLP 2018 • Yi Liao, Lidong Bing, Piji Li, Shuming Shi, Wai Lam, Tong Zhang
For example, an input sequence could be a word sequence, such as review sentence and advertisement text.
5 code implementations • 10 Oct 2018 • Jiechao Xiong, Qing Wang, Zhuoran Yang, Peng Sun, Lei Han, Yang Zheng, Haobo Fu, Tong Zhang, Ji Liu, Han Liu
Most existing deep reinforcement learning (DRL) frameworks consider either discrete action space or continuous action space solely.
no code implementations • ECCV 2018 • Yonggen Ling, Linchao Bao, Zequn Jie, Fengming Zhu, Ziyang Li, Shanmin Tang, Yongsheng Liu, Wei Liu, Tong Zhang
Our approach is able to handle the rolling-shutter effects and imperfect sensor synchronization in a unified way.
no code implementations • ECCV 2018 • Yitong Wang, Dihong Gong, Zheng Zhou, Xing Ji, Hao Wang, Zhifeng Li, Wei Liu, Tong Zhang
Extensive experiments conducted on the three public domain face aging datasets (MORPH Album 2, CACD-VS and FG-NET) have shown the effectiveness of the proposed approach and the value of the constructed CAF dataset on AIFR.
Ranked #3 on Age-Invariant Face Recognition on MORPH Album2
no code implementations • EMNLP 2018 • Baosong Yang, Zhaopeng Tu, Derek F. Wong, Fandong Meng, Lidia S. Chao, Tong Zhang
Self-attention networks have proven to be of profound value for its strength of capturing global dependencies.
Ranked #29 on Machine Translation on WMT2014 English-German
no code implementations • EMNLP 2018 • Zi-Yi Dou, Zhaopeng Tu, Xing Wang, Shuming Shi, Tong Zhang
Advanced neural machine translation (NMT) models generally implement encoder and decoder as multiple layers, which allows systems to model complex functions and capture complicated linguistic structures.
no code implementations • EMNLP 2018 • Jian Li, Zhaopeng Tu, Baosong Yang, Michael R. Lyu, Tong Zhang
Multi-head attention is appealing for the ability to jointly attend to information from different representation subspaces at different positions.
no code implementations • 2 Nov 2018 • Shixiang Chen, Shiqian Ma, Anthony Man-Cho So, Tong Zhang
We prove that the proposed method globally converges to a stationary point.
no code implementations • 2 Nov 2018 • Tong Zhang, Pan Ji, Mehrtash Harandi, Richard Hartley, Ian Reid
In this paper, we introduce a method that simultaneously learns an embedding space along subspaces within it to minimize a notion of reconstruction error, thus addressing the problem of subspace clustering in an end-to-end learning paradigm.
no code implementations • NeurIPS 2018 • Conghui Tan, Tong Zhang, Shiqian Ma, Ji Liu
Regularized empirical risk minimization problem with linear predictor appears frequently in machine learning.
no code implementations • ECCV 2018 • Kaipeng Zhang, Zhanpeng Zhang, Chia-Wen Cheng, Winston H. Hsu, Yu Qiao, Wei Liu, Tong Zhang
Face hallucination is a generative task to super-resolve the facial image with low resolution while human perception of face heavily relies on identity information.
no code implementations • 21 Nov 2018 • Xiang Kong, Zhaopeng Tu, Shuming Shi, Eduard Hovy, Tong Zhang
Although Neural Machine Translation (NMT) models have advanced state-of-the-art performance in machine translation, they face problems like the inadequate translation.
Ranked #35 on Machine Translation on WMT2014 English-German
no code implementations • 30 Nov 2018 • Keyu Yan, Wenming Zheng, Tong Zhang, Yuan Zong, Zhen Cui
Cross-database non-frontal expression recognition is a very meaningful but rather difficult subject in the fields of computer vision and affect computing.
Facial Expression Recognition Facial Expression Recognition (FER) +1
1 code implementation • NeurIPS 2018 • Qing Wang, Jiechao Xiong, Lei Han, Peng Sun, Han Liu, Tong Zhang
We consider deep policy learning with only batched historical trajectories.
no code implementations • NeurIPS 2018 • Conghui Tan, Tong Zhang, Shiqian Ma, Ji Liu
Regularized empirical risk minimization problem with linear predictor appears frequently in machine learning.
no code implementations • NeurIPS 2018 • Jianfei Chen, Jun Zhu, Yee Whye Teh, Tong Zhang
However, sEM has a slower asymptotic convergence rate than batch EM, and requires a decreasing sequence of step sizes, which is difficult to tune.
no code implementations • NeurIPS 2018 • Cong Fang, Chris Junchi Li, Zhouchen Lin, Tong Zhang
Specially, we prove that the SPIDER-SFO algorithm achieves a gradient computation cost of $\mathcal{O}\left( \min( n^{1/2} \epsilon^{-2}, \epsilon^{-3} ) \right)$ to find an $\epsilon$-approximate first-order stationary point.
no code implementations • 6 Dec 2018 • Kaiqing Zhang, Zhuoran Yang, Han Liu, Tong Zhang, Tamer Başar
This work appears to be the first finite-sample analysis for batch MARL, a step towards rigorous theoretical understanding of general MARL algorithms in the finite-sample regime.
Multi-agent Reinforcement Learning reinforcement-learning +1
no code implementations • 19 Dec 2018 • Yuan Zong, Tong Zhang, Wenming Zheng, Xiaopeng Hong, Chuangao Tang, Zhen Cui, Guoying Zhao
Cross-database micro-expression recognition (CDMER) is one of recently emerging and interesting problem in micro-expression analysis.
no code implementations • 29 Dec 2018 • Haishan Ye, Zhichao Huang, Cong Fang, Chris Junchi Li, Tong Zhang
Zeroth-order optimization is an important research topic in machine learning.
1 code implementation • 7 Jan 2019 • Baoyuan Wu, Weidong Chen, Yanbo Fan, Yong Zhang, Jinlong Hou, Jie Liu, Tong Zhang
In this work, we propose to train CNNs from images annotated with multiple tags, to enhance the quality of visual representation of the trained CNN model.
no code implementations • 1 Feb 2019 • Cong Fang, Zhouchen Lin, Tong Zhang
In this paper, we give a sharp analysis for Stochastic Gradient Descent (SGD) and prove that SGD is able to efficiently escape from saddle points and find an $(\epsilon, O(\epsilon^{0. 5}))$-approximate second-order stationary point in $\tilde{O}(\epsilon^{-3. 5})$ stochastic gradient computations for generic nonconvex optimization problems, when the objective function satisfies gradient-Lipschitz, Hessian-Lipschitz, and dispersive noise assumptions.
no code implementations • 15 Feb 2019 • Zi-Yi Dou, Zhaopeng Tu, Xing Wang, Long-Yue Wang, Shuming Shi, Tong Zhang
With the promising progress of deep neural networks, layer aggregation has been used to fuse information across layers in various fields, such as computer vision and machine translation.
no code implementations • CVPR 2019 • Yinpeng Dong, Hang Su, Baoyuan Wu, Zhifeng Li, Wei Liu, Tong Zhang, Jun Zhu
In this paper, we evaluate the robustness of state-of-the-art face recognition models in the decision-based black-box attack setting, where the attackers have no access to the model parameters and gradients, but can only acquire hard-label predictions by sending queries to the target model.
no code implementations • 24 Apr 2019 • Tong Zhang, Pan Ji, Mehrtash Harandi, Wenbing Huang, Hongdong Li
We introduce the Neural Collaborative Subspace Clustering, a neural model that discovers clusters of data points drawn from a union of low-dimensional subspaces.
1 code implementation • ICLR 2019 • Meng Fang, Cheng Zhou, Bei Shi, Boqing Gong, Jia Xu, Tong Zhang
Dealing with sparse rewards is one of the most important challenges in reinforcement learning (RL), especially when a goal is dynamic (e. g., to grasp a moving object).
1 code implementation • 1 May 2019 • Yandong Li, Lijun Li, Liqiang Wang, Tong Zhang, Boqing Gong
Powerful adversarial attack methods are vital for understanding how to construct robust deep neural networks (DNNs) and for thoroughly testing defense techniques.
no code implementations • ICLR 2019 • Yandong Li, Lijun Li, Liqiang Wang, Tong Zhang, Boqing Gong
In other words, there is a population of adversarial examples, instead of only one, for any input to a DNN.
1 code implementation • 9 May 2019 • Baoyuan Wu, Li Shen, Tong Zhang, Bernard Ghanem
Thus, LS-LP is equivalent to the original MAP inference problem.
no code implementations • 15 May 2019 • Hanlin Tang, Xiangru Lian, Chen Yu, Tong Zhang, Ji Liu
For example, under the popular parameter server model for distributed learning, the worker nodes need to send the compressed local gradients to the parameter server, which performs the aggregation.
no code implementations • 28 May 2019 • Kean Ming Tan, Junwei Lu, Tong Zhang, Han Liu
To address this issue, neuroscientists have been measuring brain activity under natural viewing experiments in which the subjects are given continuous stimuli, such as watching a movie or listening to a story.
1 code implementation • ACL 2019 • Miaofeng Liu, Yan Song, Hongbin Zou, Tong Zhang
Supervised models suffer from the problem of domain shifting where distribution mismatch in the data across domains greatly affect model performance.
1 code implementation • 11 Jul 2019 • Yujiao Shi, Xin Yu, Liu Liu, Tong Zhang, Hongdong Li
This paper proposes a novel Cross-View Feature Transport (CVFT) technique to explicitly establish cross-view domain transfer that facilitates feature alignment between ground and aerial images.
no code implementations • 17 Jul 2019 • Hanlin Tang, Xiangru Lian, Shuang Qiu, Lei Yuan, Ce Zhang, Tong Zhang, Ji Liu
Since the \emph{decentralized} training has been witnessed to be superior to the traditional \emph{centralized} training in the communication restricted scenario, therefore a natural question to ask is "how to apply the error-compensated technology to the decentralized learning to further reduce the communication cost."
no code implementations • 28 Aug 2019 • Tong Zhang, Laurence H. Jackson, Alena Uus, James R. Clough, Lisa Story, Mary A. Rutherford, Joseph V. Hajnal, Maria Deprez
The results show that the proposed pipeline can accurately estimate the respiratory state and reconstruct 4D SR volumes with better or similar performance to the 3D SVR pipeline with less than 20\% sparsely selected slices.
no code implementations • 25 Sep 2019 • Han Shi, Renjie Pi, Hang Xu, Zhenguo Li, James T. Kwok, Tong Zhang
Inspired by the nature of the graph structure of a neural network, we propose BOGCN-NAS, a NAS algorithm using Bayesian Optimization with Graph Convolutional Network (GCN) predictor.
1 code implementation • 26 Sep 2019 • Guilin Li, Xing Zhang, Zitong Wang, Matthias Tan, Jiashi Feng, Zhenguo Li, Tong Zhang
Recently, the efficiency of automatic neural architecture design has been significantly improved by gradient-based search methods such as DARTS.
no code implementations • 21 Oct 2019 • Ming-Han Yang, Andre Milzarek, Zaiwen Wen, Tong Zhang
In this paper, a novel stochastic extra-step quasi-Newton method is developed to solve a class of nonsmooth nonconvex composite optimization problems.
no code implementations • 25 Oct 2019 • Haishan Ye, Tong Zhang
We show that the estimated covariance matrix of MiNES converges to the inverse of Hessian matrix of the objective function with a sublinear convergence rate.
no code implementations • 25 Oct 2019 • Cong Fang, Hanze Dong, Tong Zhang
Recently, over-parameterized neural networks have been extensively analyzed in the literature.
7 code implementations • Findings of the Association for Computational Linguistics 2020 • Shizhe Diao, Jiaxin Bai, Yan Song, Tong Zhang, Yonggang Wang
Moreover, it is shown that reasonable performance can be obtained when ZEN is trained on a small corpus, which is important for applying pre-training techniques to scenarios with limited data.
Ranked #1 on Chinese Part-of-Speech Tagging on CTB5 Dev
Chinese Named Entity Recognition Chinese Word Segmentation +5
no code implementations • 5 Nov 2019 • Jianchen Zhu, Tong Zhang, Shengjie Zhao, Carlos Hinojosa, Zengli Liu, Gonzalo R. Arce
This paper aims at developing a clustering approach with spectral images directly from CASSI compressive measurements.
no code implementations • 7 Nov 2019 • Tong Zhang, Fatih Porikli
The residual at a layer is computed by the difference between the aggregated reconstructions of the previous layers and the downsampled original image at that layer.
1 code implementation • ICLR 2020 • Zhichao Huang, Tong Zhang
We present a new method for black-box adversarial attack.
no code implementations • 18 Nov 2019 • Cong Fang, Yihong Gu, Weizhong Zhang, Tong Zhang
This new analysis is consistent with empirical observations that deep neural networks are capable of learning efficient feature representations.
no code implementations • 20 Nov 2019 • Daniel Chao Zhou, Zhongming Jin, Tong Zhang
As an adaptive, interpretable, robust, and accurate meta-algorithm for arbitrary differentiable loss functions, gradient tree boosting is one of the most popular machine learning techniques, though the computational expensiveness severely limits its usage.
1 code implementation • NeurIPS 2020 • Han Shi, Renjie Pi, Hang Xu, Zhenguo Li, James T. Kwok, Tong Zhang
In this work, we propose BONAS (Bayesian Optimized Neural Architecture Search), a sample-based NAS framework which is accelerated using weight-sharing to evaluate multiple related architectures simultaneously.
no code implementations • 28 Nov 2019 • Zheyan Shen, Peng Cui, Tong Zhang, Kun Kuang
We consider the problem of learning linear prediction models with model misspecification bias.
no code implementations • 28 Nov 2019 • Xueya Zhang, Tong Zhang, Wenting Zhao, Zhen Cui, Jian Yang
Graph convolutional networks (GCNs) have shown the powerful ability in text structure representation and effectively facilitate the task of text classification.
1 code implementation • NeurIPS 2019 • Qing Wang, Yingru Li, Jiechao Xiong, Tong Zhang
In deep reinforcement learning, policy optimization methods need to deal with issues such as function approximation and the reuse of off-policy data.
no code implementations • 27 Dec 2019 • Haishan Ye, Shusen Wang, Zhihua Zhang, Tong Zhang
Fast matrix algorithms have become the fundamental tools of machine learning in big data era.
1 code implementation • 8 Jan 2020 • Mingyi Liu, Zhiying Tu, Tong Zhang, Tonghua Su, Zhongjie Wang
In this paper, we first examine traditional active learning strategies in a specific case of BiLstm-CRF that has widely used in named entity recognition on several typical datasets.
no code implementations • NeurIPS 2020 • Luo Luo, Haishan Ye, Zhichao Huang, Tong Zhang
We consider nonconvex-concave minimax optimization problems of the form $\min_{\bf x}\max_{\bf y\in{\mathcal Y}} f({\bf x},{\bf y})$, where $f$ is strongly-concave in $\bf y$ but possibly nonconvex in $\bf x$ and ${\mathcal Y}$ is a convex and compact set.
no code implementations • 15 Jan 2020 • Conghui Tan, Yuqiu Qian, Shiqian Ma, Tong Zhang
Dual averaging-type methods are widely used in industrial machine learning applications due to their ability to promoting solution structure (e. g., sparsity) efficiently.
no code implementations • ICLR 2020 • Chunyan Xu, Zhen Cui, Xiaobin Hong, Tong Zhang, Jian Yang, Wei Liu
In this work, we address semi-supervised classification of graph data, where the categories of those unlabeled nodes are inferred from labeled nodes as well as graph structures.
no code implementations • NeurIPS 2020 • Zixiang Chen, Yuan Cao, Quanquan Gu, Tong Zhang
In this paper, we provide a generalized neural tangent kernel analysis and show that noisy gradient descent with weight decay can still exhibit a "kernel-like" behavior.
2 code implementations • 21 Feb 2020 • Xinwei Shen, Tong Zhang, Kani Chen
This paper considers the general $f$-divergence formulation of bidirectional generative modeling, which includes VAE and BiGAN as special cases.
3 code implementations • 27 Feb 2020 • Quan Tang, Fagui Liu, Tong Zhang, Jun Jiang, Yu Zhang
The way features propagate in Fully Convolutional Networks is of momentous importance to capture multi-scale contexts for obtaining precise segmentation masks.
Ranked #23 on Semantic Segmentation on SUN-RGBD (using extra training data)
1 code implementation • CVPR 2020 • Chaoyang He, Haishan Ye, Li Shen, Tong Zhang
To remedy this, this paper proposes \mldas, a mixed-level reformulation for NAS that can be optimized efficiently and reliably.
1 code implementation • CVPR 2020 • Jing Zhang, Deng-Ping Fan, Yuchao Dai, Saeed Anwar, Fatemeh Sadat Saleh, Tong Zhang, Nick Barnes
In this paper, we propose the first framework (UCNet) to employ uncertainty for RGB-D saliency detection by learning from the data labeling process.
Ranked #4 on RGB-D Salient Object Detection on LFSD
1 code implementation • 21 Apr 2020 • Shizhe Diao, Yan Song, Tong Zhang
Keyphrase generation aims to produce a set of phrases summarizing the essentials of a given document.
1 code implementation • 25 Apr 2020 • Onur Barut, Yan Luo, Tong Zhang, Weigang Li, Peilong Li
Classifying network traffic is the basis for important network applications.
no code implementations • 2 May 2020 • Haishan Ye, Luo Luo, Ziang Zhou, Tong Zhang
This paper considers the decentralized convex optimization problem, which has a wide range of applications in large-scale machine learning, sensor networks, and control theory.
1 code implementation • CVPR 2020 • Qi Chang, Hui Qu, Yikai Zhang, Mert Sabuncu, Chao Chen, Tong Zhang, Dimitris Metaxas
In this paper, we propose a data privacy-preserving and communication efficient distributed GAN learning framework named Distributed Asynchronized Discriminator GAN (AsynDGAN).
1 code implementation • 7 Jun 2020 • Zhenbo Song, Jianfeng Lu, Tong Zhang, Hongdong Li
In this paper, we propose a monocular camera-based inter-vehicle distance and relative velocity estimation method based on end-to-end training of a deep neural network.
1 code implementation • 27 Jun 2020 • Jason Ge, Xingguo Li, Haoming Jiang, Han Liu, Tong Zhang, Mengdi Wang, Tuo Zhao
We describe a new library named picasso, which implements a unified framework of pathwise coordinate optimization for a variety of sparse learning problems (e. g., sparse linear regression, sparse logistic regression, sparse Poisson regression and scaled sparse linear regression) combined with efficient active set selection strategies.
no code implementations • ICML 2020 • Rie Johnson, Tong Zhang
This paper presents a framework of successive functional gradient optimization for training nonconvex models such as neural networks, where training is driven by mirror descent in a function space.
1 code implementation • ACL 2020 • Yuanhe Tian, Yan Song, Xiang Ao, Fei Xia, Xiaojun Quan, Tong Zhang, Yonggang Wang
Chinese word segmentation (CWS) and part-of-speech (POS) tagging are important fundamental tasks for Chinese language processing, where joint learning of them is an effective one-step solution for both tasks.
1 code implementation • ACL 2020 • Yuanhe Tian, Yan Song, Fei Xia, Tong Zhang, Yonggang Wang
Contextual features always play an important role in Chinese word segmentation (CWS).
Ranked #1 on Chinese Word Segmentation on CITYU
no code implementations • 3 Jul 2020 • Cong Fang, Jason D. Lee, Pengkun Yang, Tong Zhang
This new representation overcomes the degenerate situation where all the hidden units essentially have only one meaningful hidden unit in each middle layer, and further leads to a simpler representation of DNNs, for which the training objective can be reformulated as a convex optimization problem via suitable re-parameterization.
no code implementations • ECCV 2020 • Xin Chen, Yawen Duan, Zewei Chen, Hang Xu, Zihao Chen, Xiaodan Liang, Tong Zhang, Zhenguo Li
In spite of its remarkable progress, many algorithms are restricted to particular search spaces.
Ranked #13 on Neural Architecture Search on NAS-Bench-201, ImageNet-16-120 (Accuracy (Val) metric)
no code implementations • 26 Jul 2020 • Hongtao Yang, Tong Zhang, Wenbing Huang, Xuming He, Fatih Porikli
To be clear, in this paper, we refer unsupervised learning as learning without task-specific human annotations, pairs or any form of weak supervision.)
no code implementations • ECCV 2020 • Xueya Zhang, Tong Zhang, Xiaobin Hong, Zhen Cui, Jian Yang
Spectral graph filtering is introduced to encode graph signals, which are then embedded as probability distributions in a Wasserstein space, called graph Wasserstein metric learning.
no code implementations • 19 Aug 2020 • Yun Wang, Tong Zhang, Zhen Cui, Chunyan Xu, Jian Yang
For label diffusion of instance-awareness in graph convolution, rather than using the statistical label correlation alone, an image-dependent label correlation matrix (LCM), fusing both the statistical LCM and an individual one of each image instance, is constructed for graph inference on labels to inject adaptive information of label-awareness into the learned features of the model.
no code implementations • 24 Sep 2020 • Qianliang Wu, Tong Zhang, Zhen Cui, Jian Yang
In this paper, we aim to mine the cue of user preferences in resource-limited recommendation tasks, for which purpose we specifically build a large used car transaction dataset possessing resource-limitation characteristics.
no code implementations • 3 Oct 2020 • Zhichao Huang, Yaowei Huang, Tong Zhang
We show that searching over the structured space can be approximated by a time-varying contextual bandits problem, where the attacker takes feature of the associated arm to make modifications of the input, and receives an immediate reward as the reduction of the loss function.
1 code implementation • 6 Oct 2020 • Xinwei Shen, Furui Liu, Hanze Dong, Qing Lian, Zhitang Chen, Tong Zhang
This paper proposes a Disentangled gEnerative cAusal Representation (DEAR) learning method under appropriate supervised information.
2 code implementations • 7 Oct 2020 • Yuhui Ding, Quanming Yao, Huan Zhao, Tong Zhang
Specifically, we search for a meta graph, which can capture more complex semantic relations than a meta path, to determine how graph neural networks (GNNs) propagate messages along different types of edges.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Yuanhe Tian, Yan Song, Fei Xia, Tong Zhang
Constituency parsing is a fundamental and important task for natural language understanding, where a good representation of contextual information can help this task.
Ranked #1 on Constituency Parsing on ATB
10 code implementations • 28 Oct 2020 • Kai Han, Yunhe Wang, Qiulin Zhang, Wei zhang, Chunjing Xu, Tong Zhang
To this end, we summarize a tiny formula for downsizing neural architectures through a series of smaller models derived from the EfficientNet-B0 with the FLOPs constraint.
Ranked #692 on Image Classification on ImageNet
1 code implementation • 3 Nov 2020 • Bochao Wang, Hang Xu, Jiajin Zhang, Chen Chen, Xiaozhi Fang, Yixing Xu, Ning Kang, Lanqing Hong, Chenhan Jiang, Xinyue Cai, Jiawei Li, Fengwei Zhou, Yong Li, Zhicheng Liu, Xinghao Chen, Kai Han, Han Shu, Dehua Song, Yunhe Wang, Wei zhang, Chunjing Xu, Zhenguo Li, Wenzhi Liu, Tong Zhang
Automated Machine Learning (AutoML) is an important industrial solution for automatic discovery and deployment of the machine learning models.
no code implementations • NeurIPS 2020 • Haishan Ye, Ziang Zhou, Luo Luo, Tong Zhang
In this paper, we propose a new method which establishes the optimal computational complexity and a near optimal communication complexity.
3 code implementations • NeurIPS 2020 • Kai Han, Yunhe Wang, Qiulin Zhang, Wei zhang, Chunjing Xu, Tong Zhang
To this end, we summarize a tiny formula for downsizing neural architectures through a series of smaller models derived from the EfficientNet-B0 with the FLOPs constraint.
1 code implementation • NeurIPS 2020 • Yihong Gu, Weizhong Zhang, Cong Fang, Jason D. Lee, Tong Zhang
With the help of a new technique called {\it neural network grafting}, we demonstrate that even during the entire training process, feature distributions of differently initialized networks remain similar at each layer.
1 code implementation • NeurIPS 2020 • Guilin Li, Junlei Zhang, Yunhe Wang, Chuanjian Liu, Matthias Tan, Yunfeng Lin, Wei zhang, Jiashi Feng, Tong Zhang
In particular, we propose a novel joint-training framework to train plain CNN by leveraging the gradients of the ResNet counterpart.
no code implementations • 7 Dec 2020 • Wenqing Chen, Lulu Liu, Wentao Yang, Dong Chen, Zhengtai Liu, Yaobo Huang, Tong Zhang, Haijun Zhang, Zhonghao Liu, D. W. Shen
Utilizing angle-resolved photoemission spectroscopy and first-principles calculations, here, we demonstrate the existence of topological nodal-line states and drumheadlike surface states in centrosymmetric superconductor SnTaS2, which is a type-II superconductor with a critical transition temperature of about 3 K. The valence bands from Ta 5d orbitals and the conduction bands from Sn 5p orbitals cross each other, forming two nodal lines in the vicinity of the Fermi energy without the inclusion of spin-orbit coupling (SOC), protected by the spatial-inversion symmetry and time-reversal symmetry.
Superconductivity
no code implementations • 15 Dec 2020 • Qi Chang, Zhennan Yan, Lohendran Baskaran, Hui Qu, Yikai Zhang, Tong Zhang, Shaoting Zhang, Dimitris N. Metaxas
As deep learning technologies advance, increasingly more data is necessary to generate general and robust models for various tasks.
1 code implementation • 27 Dec 2020 • Cong Fang, Hanze Dong, Tong Zhang
Deep learning has received considerable empirical successes in recent years.
no code implementations • 30 Dec 2020 • Haishan Ye, Wei Xiong, Tong Zhang
This paper considers the decentralized composite optimization problem.
no code implementations • 1 Jan 2021 • Wenting Zhao, Yuan Fang, Zhen Cui, Tong Zhang, Jian Yang, Wei Liu
In this paper, we propose a simple yet effective graph deformer network (GDN) to fulfill anisotropic convolution filtering on graphs, analogous to the standard convolution operation on images.
no code implementations • 1 Jan 2021 • Qing Lian, LIN Yong, Tong Zhang
We consider the domain generalization problem, where the test domain differs from the training domain.
2 code implementations • 1 Jan 2021 • Yawen Duan, Xin Chen, Hang Xu, Zewei Chen, Xiaodan Liang, Tong Zhang, Zhenguo Li
While existing NAS methods mostly design architectures on one single task, algorithms that look beyond single-task search are surging to pursue a more efficient and universal solution across various tasks.
no code implementations • 1 Jan 2021 • Xiao Zhou, Weizhong Zhang, Tong Zhang
An appealing feature of ProbMask is that the amounts of weight redundancy can be learned automatically via our constraint and thus we avoid the problem of tuning pruning rates individually for different layers in a network.
no code implementations • ICCV 2021 • Yun Wang, Tong Zhang, Xueya Zhang, Zhen Cui, Yuge Huang, Pengcheng Shen, Shaoxin Li, Jian Yang
Then, a Wasserstein coupled dictionary, containing multiple pairs of counterpart graph keys with each key corresponding to one modality, is constructed for further feature learning.
no code implementations • 3 Jan 2021 • Tong Zhang, Yinfei Xu, Shuai Wang, Miaowen Wen, Rui Wang
This paper studies the problem of sum-secure degrees of freedom (SDoF) of the (M, M, N, N) multiple-input multiple-output (MIMO) interference channel with local output feedback, so as to build an information-theoretic foundation and provide practical transmission schemes for 6G-enabled vehicles-to-vehicles (V2V).
Information Theory Information Theory
no code implementations • 26 Jan 2021 • Min Yan, Guoshan Zhang, Tong Zhang, Yueming Zhang
In inference time, we design a brand-new grouping post-processing method that relates each part instance with one single human instance and groups them together to obtain the final human-level parsing result.
no code implementations • 8 Feb 2021 • Haishan Ye, Tong Zhang
This leads to a decentralized PCA algorithm called \texttt{DeEPCA}, which has a convergence rate similar to that of the centralized PCA, while achieving the best communication complexity among existing decentralized PCA algorithms.
no code implementations • 27 Feb 2021 • Wenrui Gan, Zhulin Liu, C. L. Philip Chen, Tong Zhang
In general, the main work of this paper include: (1) propose SiLa Learning, which improves the performance of common models without increasing test parameters; (2) compares SiLa with DML and proves that SiLa can improve the generalization of the model; (3) SiLa is applied to Dynamic Neural Networks, and proved that SiLa can be used for various types of network structures.
1 code implementation • 1 Mar 2021 • Yang Yang, Jiancong Chen, Ruixuan Wang, Ting Ma, Lingwei Wang, Jie Chen, Wei-Shi Zheng, Tong Zhang
Despite tremendous efforts, it is very challenging to generate a robust model to assist in the accurate quantification assessment of COVID-19 on chest CT images.
no code implementations • 10 Mar 2021 • Xuran Xu, Tong Zhang, Chunyan Xu, Zhen Cui, Jian Yang
We further extend graph convolution into tensor space and propose a tensor graph convolution network to extract more discriminating features from spatial-temporal graph data.
Ranked #1 on Traffic Prediction on SZ-Taxi
13 code implementations • CVPR 2021 • Duo Li, Jie Hu, Changhu Wang, Xiangtai Li, Qi She, Lei Zhu, Tong Zhang, Qifeng Chen
Convolution has been the core ingredient of modern neural networks, triggering the surge of deep learning in vision.
Ranked #703 on Image Classification on ImageNet
1 code implementation • CVPR 2021 • Zhichao Huang, Xintong Han, Jia Xu, Tong Zhang
We present a new method for few-shot human motion transfer that achieves realistic human image generation with only a small number of appearance inputs.
2 code implementations • CVPR 2021 • Aixuan Li, Jing Zhang, Yunqiu Lv, Bowen Liu, Tong Zhang, Yuchao Dai
Visual salient object detection (SOD) aims at finding the salient object(s) that attract human attention, while camouflaged object detection (COD) on the contrary intends to discover the camouflaged object(s) that hidden in the surrounding.
1 code implementation • 8 Apr 2021 • Bahar Aydemir, Deblina Bhattacharjee, Tong Zhang, Seungryong Kim, Mathieu Salzmann, Sabine Süsstrunk
Saliency prediction has made great strides over the past two decades, with current techniques modeling low-level information, such as color, intensity and size contrasts, and high-level ones, such as attention and gaze direction for entire objects.
no code implementations • CVPR 2021 • Jie Hong, Pengfei Fang, Weihao Li, Tong Zhang, Christian Simon, Mehrtash Harandi, Lars Petersson
Few-shot learning aims to correctly recognize query samples from unseen classes given a limited number of support samples, often by relying on global embeddings of images.
no code implementations • CVPR 2022 • Qing Lian, Botao Ye, Ruijia Xu, Weilong Yao, Tong Zhang
In addition, we demonstrate that the augmentation methods are well suited for semi-supervised training and cross-dataset generalization.
1 code implementation • CVPR 2021 • Xiao Zhou, Weizhong Zhang, Hang Xu, Tong Zhang
Weight pruning is an effective technique to reduce the model size and inference time for deep neural networks in real-world deployments.
1 code implementation • 4 May 2021 • Yan Song, Tong Zhang, Yonggang Wang, Kai-Fu Lee
Pre-trained text encoders have drawn sustaining attention in natural language processing (NLP) and shown their capability in obtaining promising results in different tasks.