no code implementations • NeurIPS 2012 • Chao Zhang, Lei Zhang, Jieping Ye
Afterwards, we analyze the asymptotic convergence and the rate of convergence of the learning process for such kind of domain adaptation.
no code implementations • 23 Apr 2013 • Hongyang Zhang, Zhouchen Lin, Chao Zhang
For several rank minimization problems, such a replacement has been theoretically proven to be valid, i. e., the solution to nuclear norm minimization problem is also the solution to rank minimization problem.
no code implementations • 26 Sep 2013 • Chao Zhang
samples, and then show that the generalization bounds have a faster rate of convergence than the traditional results.
no code implementations • 2 Jan 2014 • Chao Zhang, Lei Zhang, Wei Fan, Jieping Ye
Finally, we analyze the asymptotic convergence and the rate of convergence of the learning process for representative domain adaptation.
no code implementations • 19 May 2014 • Chao Zhang, Hong-cen Mei, Hao Yang
A large number of experimental data shows that Support Vector Machine (SVM) algorithm has obvious advantages in text classification, handwriting recognition, image classification, bioinformatics, and some other fields.
no code implementations • CVPR 2014 • Guodong Guo, Chao Zhang
Further, we study the amount of data needed in the target population to learn a cross-population age estimator.
no code implementations • 28 Aug 2014 • Chao Zhang, DaCheng Tao, Tao Hu, Xiang Li
We are mainly concerned with two theoretical questions: 1) under what conditions does RMTL perform better with a smaller task sample size than STL?
no code implementations • 6 Dec 2014 • Hongyang Zhang, Zhouchen Lin, Chao Zhang, Junbin Gao
More specifically, we discover that once a solution to one of the models is obtained, we can obtain the solutions to other models in closed-form formulations.
1 code implementation • 4 Jan 2015 • Chao Zhang, Chunhua Shen, Tingzhi Shen
We experimentally demonstrate that the learned features, together with our matching model, outperforms state-of-the-art methods such as the SIFT flow, coherency sensitive hashing and the recent deformable spatial pyramid matching methods both in terms of accuracy and computation efficiency.
no code implementations • CVPR 2015 • Junchi Yan, Chao Zhang, Hongyuan Zha, Wei Liu, Xiaokang Yang, Stephen M. Chu
Evaluations on both synthetic and real-world data corroborate the efficiency of our method.
no code implementations • 25 Jun 2015 • Hongyang Zhang, Zhouchen Lin, Chao Zhang
As an application, we also find that the solutions to extended robust Low-Rank Representation and to our extended robust MC are mutually expressible, so both our theory and algorithm can be applied to the subspace clustering problem with missing values under certain conditions.
no code implementations • ICCV 2015 • Chao Zhang, Behrend Heeren, Martin Rumpf, William A. P. Smith
In this paper we describe how to perform Principal Components Analysis in "shell space".
no code implementations • 25 May 2016 • Yanxiang Chen, Yuxing Hu, Luming Zhang, Ping Li, Chao Zhang
To remedy these problems, we develop a deep architecture to learn aesthetically-relevant visual attributes from Flickr1, which are localized by multiple textual attributes in a weakly-supervised setting.
no code implementations • CVPR 2016 • Chao Zhang, William A. P. Smith, Arnaud Dessein, Nick Pears, Hang Dai
In this paper we present a method for computing dense correspondence between a set of 3D face meshes using functional maps.
no code implementations • 22 Oct 2016 • Julie Yixuan Zhu, Chao Zhang, Huichu Zhang, Shi Zhi, Victor O. K. Li, Jiawei Han, Yu Zheng
Therefore, we present \emph{p-Causality}, a novel pattern-aided causality analysis approach that combines the strengths of \emph{pattern mining} and \emph{Bayesian learning} to efficiently and faithfully identify the \emph{ST causal pathways}.
no code implementations • 13 Nov 2016 • Zebang Shen, Hui Qian, Chao Zhang, Tengfei Zhou
Algorithms with fast convergence, small number of data access, and low per-iteration complexity are particularly favorable in the big data era, due to the demand for obtaining \emph{highly accurate solutions} to problems with \emph{a large number of samples} in \emph{ultra-high} dimensional space.
1 code implementation • ICCV 2017 • Yuhui Yuan, Kuiyuan Yang, Chao Zhang
This motivates us to ensemble a set of models with different complexities in cascaded manner and mine hard examples adaptively, a sample is judged by a series of models with increasing complexities and only updates models that consider the sample as a hard case.
Ranked #14 on Image Retrieval on SOP
no code implementations • 21 Nov 2016 • Nathan D. Cahill, Harmeet Singh, Chao Zhang, Daryl A. Corcoran, Alison M. Prengaman, Paul S. Wenger, John F. Hamilton, Peter Bajorski, Andrew M. Michael
Functional connectivity analysis yields powerful insights into our understanding of the human brain.
1 code implementation • 2 Mar 2017 • Felipe Petroski Such, Shagan Sah, Miguel Dominguez, Suhas Pillai, Chao Zhang, Andrew Michael, Nathan Cahill, Raymond Ptucha
Graph-CNNs can handle both heterogeneous and homogeneous graph data, including graphs having entirely different vertex or edge sets.
1 code implementation • CVPR 2017 • Chao Zhang, Sergi Pujades, Michael Black, Gerard Pons-Moll
We address the problem of estimating human pose and body shape from 3D scans over time.
no code implementations • ICLR 2018 • Yuhui Yuan, Kuiyuan Yang, Chao Zhang
Thus, we propose feature incay to also regularize representation learning, which favors feature vectors with large norm when the samples can be correctly classified.
no code implementations • 1 Aug 2017 • Carl Yang, Lanxiao Bai, Chao Zhang, Quan Yuan, J. Han profile
In this work, we propose to devise a general and principled SSL (semi-supervised learning) framework, to alleviate data scarcity via smoothing among neighboring users and POIs, and treat various context by regularizing user preference based on context graphs.
no code implementations • 9 Sep 2017 • Shuochao Yao, Yiran Zhao, Huajie Shao, Aston Zhang, Chao Zhang, Shen Li, Tarek Abdelzaher
Recent advances in deep learning have led various applications to unprecedented achievements, which could potentially bring higher intelligence to a broad spectrum of mobile and ubiquitous applications.
1 code implementation • 23 Oct 2017 • Kai Han, Yunhe Wang, Chao Zhang, Chao Li, Chao Xu
High-dimensional data in many areas such as computer vision and machine learning tasks brings in computational and analytical difficulty.
no code implementations • 22 Jan 2018 • Sibo Yang, Chao Zhang, Wei Wu
Considered in this short note is the design of output layer nodes of feedforward neural networks for solving multi-class classification problems with r (bigger than or equal to 3) classes of samples.
no code implementations • ICLR 2019 • Yihan Gao, Chao Zhang, Jian Peng, Aditya Parameswaran
Both theoretical and empirical evidence are provided to support this argument: (a) we prove that the generalization error of these methods can be bounded by limiting the norm of vectors, regardless of the embedding dimension; (b) we show that the generalization performance of linear graph embedding methods is correlated with the norm of embedding vectors, which is small due to the early stopping of SGD and the vanishing gradients.
no code implementations • 18 Feb 2018 • Florian Kreyssig, Chao Zhang, Philip Woodland
Time delay neural networks (TDNNs) are an effective acoustic model for large vocabulary speech recognition.
no code implementations • 22 Feb 2018 • Chao Zhang, Philip Woodland
Vanishing long-term gradients are a major issue in training standard recurrent neural networks (RNNs), which can be alleviated by long short-term memory (LSTM) models with memory cells.
no code implementations • 8 May 2018 • Chao Zhang, Ce Zhu, Jimin Xiao, Xun Xu, Yipeng Liu
Finally we demonstrate the effectiveness of both approaches by visualizing the Class Activation Map (CAM) and discover that grid dropout is more aware of the whole facial areas and more robust than neuron dropout for small training dataset.
no code implementations • 18 Jun 2018 • Chao Zhang, Philip Woodland
Gating is a key technique used for integrating information from multiple sources by long short-term memory (LSTM) models and has recently also been applied to other models such as the highway network.
1 code implementation • ECCV 2018 • Yikang Li, Wanli Ouyang, Bolei Zhou, Jianping Shi, Chao Zhang, Xiaogang Wang
Generating scene graph to describe all the relations inside an image gains increasing interests these years.
Ranked #1 on Scene Graph Generation on VRD
1 code implementation • 10 Jul 2018 • Yu Shi, Qi Zhu, Fang Guo, Chao Zhang, Jiawei Han
To cope with the challenges in the comprehensive transcription of HINs, we propose the HEER algorithm, which embeds HINs via edge representations that are further coupled with properly-learned heterogeneous metrics.
1 code implementation • 2 Sep 2018 • Yu Meng, Jiaming Shen, Chao Zhang, Jiawei Han
Although many semi-supervised and weakly-supervised text classification models exist, they cannot be easily applied to deep neural models and meanwhile support limited supervision types.
8 code implementations • 4 Sep 2018 • Yuhui Yuan, Lang Huang, Jianyuan Guo, Chao Zhang, Xilin Chen, Jingdong Wang
To capture richer context information, we further combine our interlaced sparse self-attention scheme with the conventional multi-scale context schemes including pyramid pooling~\citep{zhao2017pyramid} and atrous spatial pyramid pooling~\citep{chen2018deeplab}.
Ranked #9 on Semantic Segmentation on Trans10K
no code implementations • 5 Sep 2018 • Chao Zhang, Xuequan Lu, Takuya Akashi
To settle this issue, we propose a blur-countering method for detecting valid keypoints for various types and degrees of blurred images.
no code implementations • NeurIPS 2018 • Zhisheng Zhong, Tiancheng Shen, Yibo Yang, Zhouchen Lin, Chao Zhang
To solve these problems, we propose the Super-Resolution CliqueNet (SRCliqueNet) to reconstruct the high resolution (HR) image with better textural details in the wavelet domain.
no code implementations • NeurIPS 2018 • Yiwen Guo, Chao Zhang, Chang-Shui Zhang, Yurong Chen
Deep neural networks (DNNs) are computationally/memory-intensive and vulnerable to adversarial attacks, making them prohibitive in some real-world applications.
no code implementations • 11 Nov 2018 • Chao Zhang, Min-Hsiu Hsieh, DaCheng Tao
First, we prove that the complexity of function classes convolving with vicinal functions can be controlled by that of the original function classes under the assumption that the function class is composed of Lipschitz-continuous functions.
10 code implementations • 12 Nov 2018 • Ling Zhao, Yujiao Song, Chao Zhang, Yu Liu, Pu Wang, Tao Lin, Min Deng, Haifeng Li
However, traffic forecasting has always been considered an open scientific issue, owing to the constraints of urban road network topological structure and the law of dynamic change with time, namely, spatial dependence and temporal dependence.
Ranked #4 on Traffic Prediction on SZ-Taxi
no code implementations • 26 Nov 2018 • Bold Naranchimeg, Chao Zhang, Takuya Akashi
In recent decade, many state-of-the-art algorithms on image classification as well as audio classification have achieved noticeable successes with the development of deep convolutional neural network (CNN).
Audio Classification Bird Species Classification With Audio-Visual Data +3
2 code implementations • NeurIPS 2018 • Shupeng Su, Chao Zhang, Kai Han, Yonghong Tian
To convert the input into binary code, hashing algorithm has been widely used for approximate nearest neighbor search on large-scale image sets due to its computation and storage efficiency.
2 code implementations • 22 Dec 2018 • Chao Zhang, Fangbo Tao, Xiusi Chen, Jiaming Shen, Meng Jiang, Brian Sadler, Michelle Vanni, Jiawei Han
Our method, TaxoGen, uses term embeddings and hierarchical clustering to construct a topic taxonomy in a recursive fashion.
Databases
1 code implementation • 29 Dec 2018 • Yu Meng, Jiaming Shen, Chao Zhang, Jiawei Han
During the training process, our model features a hierarchical neural structure, which mimics the given hierarchy and is capable of determining the proper levels for documents with a blocking mechanism.
1 code implementation • 2 Jan 2019 • Kai Han, Jianyuan Guo, Chao Zhang, Mingjian Zhu
Based on the considerations above, we propose a novel Attribute-Aware Attention Model ($A^3M$), which can learn local attribute representation and global category representation simultaneously in an end-to-end manner.
Ranked #4 on Fine-Grained Image Classification on CompCars
no code implementations • 8 Feb 2019 • Guangzhi Sun, Chao Zhang, Phil Woodland
This combination uses a 2-dimensional (2D) self-attentive structure, which extends the standard self-attentive layer by averaging not only across time but also across different types of embeddings.
1 code implementation • 17 Mar 2019 • Xiangpan Ji, Wenqiang Gu, Xin Qian, Hanyu Wei, Chao Zhang
We describe an approximation to the widely-used Poisson-likelihood chi-square using a linear combination of Neyman's and Pearson's chi-squares, namely "combined Neyman-Pearson chi-square" ($\chi^2_{\mathrm{CNP}}$).
Data Analysis, Statistics and Probability High Energy Physics - Experiment Nuclear Experiment
no code implementations • 26 Mar 2019 • Chunzhi Gu, Xuequan Lu, Ying He, Chao Zhang
Complex blur such as the mixup of space-variant and space-invariant blur, which is hard to model mathematically, widely exists in real images.
no code implementations • 26 Mar 2019 • Takumi Nakane, Takuya Akashi, Xuequan Lu, Chao Zhang
We propose a novel genetic algorithm to solve the image deformation estimation problem by preserving the genetic diversity.
no code implementations • 2 Apr 2019 • Hang Zou, Chao Zhang, Samson Lasaulce
The proposed point of view is fully relevant for a receiver which has to send a quantized version of the channel state to the transmitter.
1 code implementation • CVPR 2019 • Chao Zhang, Shuaicheng Liu, Xun Xu, Ce Zhu
Recently, MobileNets and ShuffleNets have been proposed to reduce the number of parameters, yielding lightweight models.
Ranked #3 on Age Estimation on FGNET
no code implementations • 11 Apr 2019 • Zhuo Lei, Chao Zhang, Qian Zhang, Guoping Qiu
In constructing the dataset, because of the subjectivity of user-generated video summarization, we manually annotate 25 summaries for each video, which are in total 1300 summaries.
no code implementations • 17 May 2019 • Hang Zou, Chao Zhang, Samson Lasaulce, Lucas Saludjian, Patrick Panciatici
In this paper, we introduce the problem of decision-oriented communications, that is, the goal of the source is to send the right amount of information in order for the intended destination to execute a task.
no code implementations • 6 Jun 2019 • Chuang-Shi Shen, Chao Zhang, Xiaosheng Gao, Yulong Li
We recognize that the fragmentation problem in shells is analogous to the cracking behavior of tree bark, and closed form solutions is obtained to describe the relationship between the expansion velocity and the number of necks with consideration of the strain rate dependent strength of the shell material.
Soft Condensed Matter Applied Physics
no code implementations • 8 Jun 2019 • Yu-cheng Chen, Matus Telgarsky, Chao Zhang, Bolton Bailey, Daniel Hsu, Jian Peng
This paper provides a simple procedure to fit generative networks to target distributions, with the goal of a small Wasserstein distance (or other optimal transport costs).
no code implementations • 18 Jun 2019 • Zhisheng Zhong, Fangyin Wei, Zhouchen Lin, Chao Zhang
Furthermore, we propose that weight tensors in networks with proper order and balanced dimension are easier to be compressed.
no code implementations • 21 Jun 2019 • Patrick von Platen, Chao Zhang, Philip Woodland
This paper proposes a novel multi-span structure for acoustic modelling based on the raw waveform with multiple streams of CNN input layers, each processing a different span of the raw waveform signal.
no code implementations • 23 Jun 2019 • Jian-Ya Ding, Chao Zhang, Lei Shen, Shengyin Li, Bing Wang, Yinghui Xu, Le Song
In many applications, a similar MIP model is solved on a regular basis, maintaining remarkable similarities in model structures and solution appearances but differing in formulation coefficients.
no code implementations • 2 Jul 2019 • Yi Zhang, Chao Zhang, Takuya Akashi
We propose a novel multi-scale template matching method which is robust against both scaling and rotation in unconstrained environments.
6 code implementations • 29 Jul 2019 • Lang Huang, Yuhui Yuan, Jianyuan Guo, Chao Zhang, Xilin Chen, Jingdong Wang
There are two successive attention modules each estimating a sparse affinity matrix.
1 code implementation • ICCV 2019 • Chao Zhang, Stephan Liwicki, William Smith, Roberto Cipolla
For the spherical domain, several methods recently adopt an icosahedron mesh, but systems are typically rotation invariant or require significant memory and parameters, thus enabling execution only at very low resolutions.
Ranked #21 on Semantic Segmentation on Stanford2D3D Panoramic
1 code implementation • 20 Aug 2019 • Yu Meng, Jiaxin Huang, Guangyuan Wang, Zihan Wang, Chao Zhang, Yu Zhang, Jiawei Han
We propose a new task, discriminative topic mining, which leverages a set of user-provided category names to mine discriminative topics from text corpora.
no code implementations • 14 Sep 2019 • Qiujia Li, Chao Zhang, Philip C. Woodland
This paper proposes a novel automatic speech recognition (ASR) framework called Integrated Source-Channel and Attention (ISCA) that combines the advantages of traditional systems based on the noisy source-channel model (SC) and end-to-end style systems using attention-based sequence-to-sequence models.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
no code implementations • 16 Sep 2019 • Hang Zou, Chao Zhang, Samson Lasaulce, Lucas Saludjian, Patrick Panciatici
We propose a framework to find a good (finite) decision set which induces a minimal performance loss w. r. t.
no code implementations • 30 Sep 2019 • Yuan-Yuan Zhao, Chao Zhang, Shuming Cheng, Xinhui Li, Yu Guo, Bi-Heng Liu, Huan-Yu Ku, Shin-Liang Chen, Qiaoyan Wen, Yun-Feng Huang, Guo-Yong Xiang, Chuan-Feng Li, Guang-Can Guo
We first establish the DI verification framework, relying on the measurement-device-independent technique and self-testing, and show it is able to verify all EPR-steerable states.
Quantum Physics
2 code implementations • 8 Oct 2019 • Yunqi Shao, Matti Hellström, Pavlin D. Mitev, Lisanne Knijff, Chao Zhang
Atomic neural networks (ANNs) constitute a class of machine learning methods for predicting potential energy surfaces and physico-chemical properties of molecules and materials.
Computational Physics Disordered Systems and Neural Networks Chemical Physics
no code implementations • 8 Oct 2019 • Chao Zhang, Min-Hsiu Hsieh, DaCheng Tao
We also develop the tail inequalities for matrix random series and matrix martingale difference sequence.
1 code implementation • 10 Oct 2019 • Wanzheng Zhu, Hongyu Gong, Jiaming Shen, Chao Zhang, Jingbo Shang, Suma Bhat, Jiawei Han
In this paper, we study the task of multi-faceted set expansion, which aims to capture all semantic facets in the seed set and return multiple sets of entities, one for each semantic facet.
no code implementations • 17 Oct 2019 • Jiaming Shen, Zeqiu Wu, Dongming Lei, Chao Zhang, Xiang Ren, Michelle T. Vanni, Brian M. Sadler, Jiawei Han
Taxonomies are of great value to many knowledge-rich applications.
no code implementations • 21 Oct 2019 • Jiahao Xie, Zebang Shen, Chao Zhang, Boyu Wang, Hui Qian
This paper focuses on projection-free methods for solving smooth Online Convex Optimization (OCO) problems.
no code implementations • 21 Oct 2019 • Chao Zhang, Jiahao Xie, Zebang Shen, Peilin Zhao, Tengfei Zhou, Hui Qian
In this paper, we explore a general Aggregated Gradient Langevin Dynamics framework (AGLD) for the Markov Chain Monte Carlo (MCMC) sampling.
1 code implementation • 22 Oct 2019 • Qiujia Li, Florian L. Kreyssig, Chao Zhang, Philip C. Woodland
In this paper, we propose Discriminative Neural Clustering (DNC) that formulates data clustering with a maximum number of clusters as a supervised sequence-to-sequence learning problem.
1 code implementation • ICCV 2019 • Jianyuan Guo, Yuhui Yuan, Lang Huang, Chao Zhang, Jinge Yao, Kai Han
On the other hand, there still exist many useful contextual cues that do not fall into the scope of predefined human parts or attributes.
Ranked #59 on Person Re-Identification on DukeMTMC-reID
1 code implementation • NeurIPS 2019 • Yu Meng, Jiaxin Huang, Guangyuan Wang, Chao Zhang, Honglei Zhuang, Lance Kaplan, Jiawei Han
While text embeddings are typically learned in the Euclidean space, directional similarity is often more effective in tasks such as word similarity and document clustering, which creates a gap between the training stage and usage stage of text embedding.
no code implementations • 10 Nov 2019 • Yassir Fathullah, Chao Zhang, Philip C. Woodland
Speaker diarisation systems nowadays use embeddings generated from speech segments in a bottleneck layer, which are needed to be discriminative for unseen speakers.
no code implementations • 10 Nov 2019 • Chao Zhang, Zichao Yang, Xiaodong He, Li Deng
This review provides a comprehensive analysis of recent works on multimodal deep learning from three perspectives: learning multimodal representations, fusing multimodal signals at various levels, and multimodal applications.
no code implementations • 20 Jan 2020 • Chao Zhang, Xuequan Lu, Katsuya Hotta, Xi Yang
The WA data can be naturally obtained in an interactive way for specific tasks, for example, in the case of homography estimation, one can easily annotate points on the same plane/object with a single label by observing the image.
4 code implementations • NeurIPS 2020 • Lang Huang, Chao Zhang, Hongyang Zhang
We propose self-adaptive training---a new training algorithm that dynamically corrects problematic training labels by model predictions without incurring extra computational cost---to improve generalization of deep learning for potentially corrupted training data.
1 code implementation • CVPR 2020 • Jianyuan Guo, Kai Han, Yunhe Wang, Chao Zhang, Zhaohui Yang, Han Wu, Xinghao Chen, Chang Xu
To this end, we propose a hierarchical trinity search framework to simultaneously discover efficient architectures for all components (i. e. backbone, neck, and head) of object detector in an end-to-end manner.
no code implementations • 30 Mar 2020 • Takumi Nakane, Xuequan Lu, Chao Zhang
In evolutionary algorithms, genetic operators iteratively generate new offspring which constitute a potentially valuable set of search history.
no code implementations • 13 Apr 2020 • Huajie Shao, Dachun Sun, Jiahao Wu, Zecheng Zhang, Aston Zhang, Shuochao Yao, Shengzhong Liu, Tianshi Wang, Chao Zhang, Tarek Abdelzaher
Motivated by this trend, we describe a novel item-item cross-platform recommender system, $\textit{paper2repo}$, that recommends relevant repositories on GitHub that match a given paper in an academic search system such as Microsoft Academic.
no code implementations • 13 Apr 2020 • Chao Zhang, Xiang Zhao, Kai Lin, Shaojun Zhang, Wen Zhao, Anzhong Wang
In particular, we find that, out of the five non-trivial field equations, only three are independent, so the problem is well-posed, as now generically there are only three unknown functions, {$F(r), B(r), A(r)$, where $F$ and $B$ are metric coefficients, and $A$ describes the aether field.}
General Relativity and Quantum Cosmology Astrophysics of Galaxies High Energy Physics - Phenomenology High Energy Physics - Theory
no code implementations • 1 May 2020 • Shi Zhi, Liyuan Liu, Yu Zhang, Shiyin Wang, Qi Li, Chao Zhang, Jiawei Han
While typical named entity recognition (NER) models require the training set to be annotated with all target types, each available datasets may only cover a part of them.
1 code implementation • 16 May 2020 • Nick Altieri, Rebecca L. Barter, James Duncan, Raaz Dwivedi, Karl Kumbier, Xiao Li, Robert Netzorg, Briton Park, Chandan Singh, Yan Shuo Tan, Tiffany Tang, Yu Wang, Chao Zhang, Bin Yu
We use this data to develop predictions and corresponding prediction intervals for the short-term trajectory of COVID-19 cumulative death counts at the county-level in the United States up to two weeks ahead.
1 code implementation • 18 Jun 2020 • Yue Yu, Yinghao Li, Jiaming Shen, Hao Feng, Jimeng Sun, Chao Zhang
We propose a self-supervised taxonomy expansion model named STEAM, which leverages natural supervision in the existing taxonomy for expansion.
1 code implementation • 28 Jun 2020 • Chen Liang, Yue Yu, Haoming Jiang, Siawpeng Er, Ruijia Wang, Tuo Zhao, Chao Zhang
We study the open-domain named entity recognition (NER) problem under distant supervision.
no code implementations • 11 Jul 2020 • Dongbo Zhang, Zheng Fang, Xuequan Lu, Hong Qin, Antonio Robles-Kelly, Chao Zhang, Ying He
3D human segmentation has seen noticeable progress in re-cent years.
1 code implementation • 18 Jul 2020 • Yu Meng, Yunyi Zhang, Jiaxin Huang, Yu Zhang, Chao Zhang, Jiawei Han
Mining a set of meaningful topics organized into a hierarchy is intuitively appealing since topic correlations are ubiquitous in massive text corpora.
Ranked #1 on Topic Models on Arxiv HEP-TH citation graph
no code implementations • 28 Jul 2020 • Wei Xue, Gang Quan, Chao Zhang, Guohong Ding, Xiaodong He, BoWen Zhou
Statistical signal processing based speech enhancement methods adopt expert knowledge to design the statistical models and linear filters, which is complementary to the deep neural network (DNN) based methods which are data-driven.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
no code implementations • 22 Aug 2020 • Jinfeng Zeng, Chenfeng Cao, Chao Zhang, Pengxiang Xu, Bei Zeng
To obtain the full spectrum of the Hamiltonian, we use a quantum imaginary time evolution algorithm with high temperature, which prepares a thermal state with a small correlation length.
Quantum Physics
2 code implementations • ICML 2020 • Lingkai Kong, Jimeng Sun, Chao Zhang
We propose a new method for quantifying uncertainties of DNNs from a dynamical system perspective.
no code implementations • 31 Aug 2020 • Chunzhi Gu, Xuequan Lu, Chao Zhang
In particular, we relate the transferred image with the example image under the Gaussian Mixture Model (GMM) and regard the transferred image color as the GMM centroids.
1 code implementation • NeurIPS 2021 • Qi Zhu, Carl Yang, Yidan Xu, Haonan Wang, Chao Zhang, Jiawei Han
Graph neural networks (GNNs) have achieved superior performance in various applications, but training dedicated GNNs can be costly for large-scale graphs.
no code implementations • 12 Sep 2020 • Yi Zhou, Shuyang Sun, Chao Zhang, Yikang Li, Wanli Ouyang
By assigning each relationship a single label, current approaches formulate the relationship detection as a classification problem.
1 code implementation • 4 Oct 2020 • Yue Yu, Kexin Huang, Chao Zhang, Lucas M. Glass, Jimeng Sun, Cao Xiao
Furthermore, most previous works focus on binary DDI prediction whereas the multi-typed DDI pharmacological effect prediction is a more meaningful but harder task.
1 code implementation • EMNLP 2020 • Rongzhi Zhang, Yue Yu, Chao Zhang
Our method, SeqMix, simply augments the queried samples by generating extra labeled sequences in each iteration.
1 code implementation • 5 Oct 2020 • Wanzheng Zhu, Chao Zhang, Shuochao Yao, Xiaobin Gao, Jiawei Han
We propose SHMM, a multi-modal spherical hidden Markov model for semantics-rich human mobility modeling.
1 code implementation • 5 Oct 2020 • Yinghao Li, Rui Feng, Isaac Rehg, Chao Zhang
We study the problem of using (partial) constituency parse trees as syntactic guidance for controlled text generation.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Wendi Ren, Yinghao Li, Hanting Su, David Kartchner, Cassie Mitchell, Chao Zhang
We study the problem of learning neural text classifiers without using any labeled data, but only easy-to-provide rules as multiple weak supervision sources.
2 code implementations • EMNLP 2020 • Yu Meng, Yunyi Zhang, Jiaxin Huang, Chenyan Xiong, Heng Ji, Chao Zhang, Jiawei Han
In this paper, we explore the potential of only using the label name of each class to train classification models on unlabeled data, without using any labeled documents.
1 code implementation • NAACL 2021 • Yue Yu, Simiao Zuo, Haoming Jiang, Wendi Ren, Tuo Zhao, Chao Zhang
To address this problem, we develop a contrastive self-training framework, COSINE, to enable fine-tuning LMs with weak supervision.
Ranked #1 on Word Sense Disambiguation on Words in Context
no code implementations • 19 Oct 2020 • Hao Wang, Jia Zhang, Yingce Xia, Jiang Bian, Chao Zhang, Tie-Yan Liu
However, most existing studies overlook the code's intrinsic structural logic, which indeed contains a wealth of semantic information, and fails to capture intrinsic features of codes.
no code implementations • 21 Oct 2020 • Rui Feng, Jie Yuan, Chao Zhang
We argue that the event extraction models so trained are inherently label-hungry, and can generalize poorly across domains and text genres. We propose a reading comprehension framework for event extraction. Specifically, we formulate event detection as a textual entailment prediction problem, and argument detection as a question answer-ing problem.
1 code implementation • EMNLP 2020 • Lingkai Kong, Haoming Jiang, Yuchen Zhuang, Jie Lyu, Tuo Zhao, Chao Zhang
Fine-tuned pre-trained language models can suffer from severe miscalibration for both in-distribution and out-of-distribution (OOD) data due to over-parameterization.
no code implementations • 22 Oct 2020 • Guangzhi Sun, Chao Zhang, Phil Woodland
Significant progress has recently been made in speaker diarisation after the introduction of d-vectors as speaker embeddings extracted from neural network (NN) speaker classifiers for clustering speech segments.
no code implementations • 27 Oct 2020 • Wen Wu, Chao Zhang, Philip C. Woodland
In this paper, a novel two-branch neural network model structure is proposed for multimodal emotion recognition, which consists of a time synchronous branch (TSB) and a time asynchronous branch (TAB).
Automatic Speech Recognition Automatic Speech Recognition (ASR) +6
no code implementations • 6 Nov 2020 • Guanghui Xu, Wei Song, Zhengchen Zhang, Chao Zhang, Xiaodong He, BoWen Zhou
Despite prosody is related to the linguistic information up to the discourse structure, most text-to-speech (TTS) systems only take into account that within each sentence, which makes it challenging when converting a paragraph of texts into natural and expressive speech.
1 code implementation • 25 Nov 2020 • Wei Wang, Chao Zhang, Xiaopei Wu
Accent recognition with deep learning framework is a similar work to deep speaker identification, they're both expected to give the input speech an identifiable representation.
no code implementations • 2 Dec 2020 • Weijie Liu, Chao Zhang, Jiahao Xie, Zebang Shen, Hui Qian, Nenggan Zheng
Graph matching finds the correspondence of nodes across two graphs and is a basic task in graph-based machine learning.
no code implementations • 22 Dec 2020 • Samuel Yen-Chi Chen, Tzu-Chieh Wei, Chao Zhang, Haiwang Yu, Shinjae Yoo
This work presents a quantum convolutional neural network (QCNN) for the classification of high energy physics events.
no code implementations • 5 Jan 2021 • Chao Zhang, Joaquin Vanschoren, Arlette van Wissen, Daniel Lakens, Boris de Ruyter, Wijnand A. IJsselsteijn
Psychological theories of habit posit that when a strong habit is formed through behavioral repetition, it can trigger behavior automatically in the same environment.
no code implementations • 15 Jan 2021 • Samuel Yen-Chi Chen, Tzu-Chieh Wei, Chao Zhang, Haiwang Yu, Shinjae Yoo
This research provides a hybrid quantum-classical graph convolutional network (QGCNN) for learning HEP data.
2 code implementations • 21 Jan 2021 • Lang Huang, Chao Zhang, Hongyang Zhang
We propose self-adaptive training -- a unified training algorithm that dynamically calibrates and enhances training processes by model predictions without incurring an extra computational cost -- to advance both supervised and self-supervised learning of deep neural networks.
no code implementations • 3 Mar 2021 • Chao Zhang, Wenqiang Yi, Yuanwei Liu, Qiang Wang
Numerical results indicate that 1) although the interference from other cells is enhanced via the RISs, the performance of the RIS-aided user still enhances since the channel quality is strengthened more obviously; and 2) the SIC order can be altered by employing the RISs since the RISs improve the channel quality of the aided user.
Information Theory Information Theory
no code implementations • 10 Mar 2021 • Chao Zhang, Shihan Wang, Henk Aarts, Mehdi Dastani
Reinforcement learning (RL) agents in human-computer interactions applications require repeated user interactions before they can perform well.
no code implementations • 12 Mar 2021 • Adnan Haider, Chao Zhang, Florian L. Kreyssig, Philip C. Woodland
This paper presents a novel natural gradient and Hessian-free (NGHF) optimisation framework for neural network training that can operate efficiently in a distributed manner.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
1 code implementation • CVPR 2021 • Yingjie Cai, Xuesong Chen, Chao Zhang, Kwan-Yee Lin, Xiaogang Wang, Hongsheng Li
The key insight is that we decouple the instances from a coarsely completed semantic scene instead of a raw input image to guide the reconstruction of instances and the overall scene.
Ranked #1 on 3D Semantic Scene Completion on NYUv2
no code implementations • 9 Apr 2021 • Chao Zhang, Xiaojun Chen, Shiqian Ma
In this paper, we propose a Riemannian smoothing steepest descent method to minimize a nonconvex and non-Lipschitz function on submanifolds.
no code implementations • 12 Apr 2021 • Xianjie Gao, Xueguan Song, Maolin Shi, Chao Zhang, Hongwei Zhang
In this paper, based on in-situ TBM operational data, we use the machine-learning (ML) methods to build the real-time forecast models for TBM load parameters, which can instantaneously provide the future values of the TBM load parameters as long as the current data are collected.
no code implementations • 12 Apr 2021 • Samson Lasaulce, Chao Zhang, Vineeth Varma, Irinel Constantin Morarescu
Should the measures be more (or less) restrictive?
no code implementations • 23 Apr 2021 • Sheldon Fung, Xuequan Lu, Chao Zhang, Chang-Tsun Li
Extensive experiments show that our unsupervised learning method enables comparable detection performance to state-of-the-art supervised techniques, in both the intra- and inter-dataset settings.
no code implementations • 23 Apr 2021 • Yi He, Haoran Xie, Chao Zhang, Xi Yang, Kazunori Miyata
This paper proposes a deep generative model for generating normal maps from users sketch with geometric sampling.
4 code implementations • 26 Apr 2021 • Wei Zeng, Xiaozhe Ren, Teng Su, Hui Wang, Yi Liao, Zhiwei Wang, Xin Jiang, ZhenZhang Yang, Kaisheng Wang, Xiaoda Zhang, Chen Li, Ziyan Gong, Yifan Yao, Xinjing Huang, Jun Wang, Jianfeng Yu, Qi Guo, Yue Yu, Yan Zhang, Jin Wang, Hengtao Tao, Dasen Yan, Zexuan Yi, Fang Peng, Fangqing Jiang, Han Zhang, Lingfeng Deng, Yehong Zhang, Zhe Lin, Chao Zhang, Shaojie Zhang, Mingyue Guo, Shanzhi Gu, Gaojun Fan, YaoWei Wang, Xuefeng Jin, Qun Liu, Yonghong Tian
To enhance the generalization ability of PanGu-$\alpha$, we collect 1. 1TB high-quality Chinese data from a wide range of domains to pretrain the model.
Ranked #1 on Reading Comprehension (One-Shot) on DuReader
Cloze (multi-choices) (Few-Shot) Cloze (multi-choices) (One-Shot) +19
1 code implementation • 8 May 2021 • Lingwei Peng, Hui Qian, Zhebang Shen, Chao Zhang, Fei Li
Model-free deep reinforcement learning has achieved great success in many domains, such as video games, recommendation systems and robotic control tasks.
2 code implementations • ACL 2021 • Yinghao Li, Pranav Shetty, Lucas Liu, Chao Zhang, Le Song
To address this challenge, we propose a conditional hidden Markov model (CHMM), which can effectively infer true labels from multi-source noisy labels in an unsupervised way.
1 code implementation • 29 May 2021 • Jiahao Xie, Chao Zhang, Zebang Shen, Weijie Liu, Hui Qian
We establish theoretical guarantees of CDMA under different choices of hyperparameters and conduct experiments on AUC maximization, robust adversarial network training, and GAN training tasks.
no code implementations • 2 Jun 2021 • Chao Zhang, Samson Lasaulce, Martin Hennebel, Lucas Saludjian, Patrick Panciatici, H. Vincent Poor
For this purpose, we formulate the framework of decision-making oriented clustering and propose an algorithm providing a decision-based partition of the data space and good representative decisions.
1 code implementation • NeurIPS 2021 • Harshavardhan Kamarthi, Lingkai Kong, Alexander Rodríguez, Chao Zhang, B. Aditya Prakash
We model the forecasting task as a probabilistic generative process and propose a functional neural process model called EPIFNP, which directly models the probability density of the forecast value.
no code implementations • 8 Jun 2021 • Takumi Nakane, Haoran Xie, Chao Zhang
Specifically, by partitioning the template image into several regions and measuring the similarity of each region independently, multiple objectives are built and deformation estimation can thus be realized by solving the MOP with off-the-shelf multi-objective evolutionary algorithms (MOEAs).
no code implementations • ICML Workshop AML 2021 • Fangcheng Liu, Chao Zhang, Hongyang Zhang
In this work, we propose a \emph{geometry-aware framework} to generate transferable adversarial perturbation with minimum norm for each input.
no code implementations • CVPR 2021 • Jianyuan Guo, Kai Han, Han Wu, Chao Zhang, Xinghao Chen, Chunjing Xu, Chang Xu, Yunhe Wang
In this paper, we present a positive-unlabeled learning based scheme to expand training data by purifying valuable images from massive unlabeled ones, where the original training data are viewed as positive data and the unlabeled images in the wild are unlabeled data.
no code implementations • 24 Jun 2021 • Lixue Liu, Chao Zhang, DaCheng Tao
Multi-fidelity data fusion (MDF) methods aims to use massive LF samples and small amounts of HF samples to develop an accurate and efficient model for describing the system with a reasonable computation burden.
no code implementations • 29 Jun 2021 • Fadi Boutros, Naser Damer, Jan Niklas Kolf, Kiran Raja, Florian Kirchbuchner, Raghavendra Ramachandra, Arjan Kuijper, Pengcheng Fang, Chao Zhang, Fei Wang, David Montero, Naiara Aginako, Basilio Sierra, Marcos Nieto, Mustafa Ekrem Erakin, Ugur Demir, Hazim Kemal, Ekenel, Asaki Kataoka, Kohei Ichikawa, Shizuma Kubo, Jie Zhang, Mingjie He, Dan Han, Shiguang Shan, Klemen Grm, Vitomir Štruc, Sachith Seneviratne, Nuran Kasthuriarachchi, Sanka Rasnayaka, Pedro C. Neto, Ana F. Sequeira, Joao Ribeiro Pinto, Mohsen Saffari, Jaime S. Cardoso
These teams successfully submitted 18 valid solutions.
1 code implementation • 1 Jul 2021 • Qiujia Li, Chao Zhang, Philip C. Woodland
Commonly used automatic speech recognition (ASR) systems can be classified into frame-synchronous and label-synchronous categories, based on whether the speech is decoded on a per-frame or per-label basis.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
1 code implementation • 3 Jul 2021 • Zhiwei Hao, Jianyuan Guo, Ding Jia, Kai Han, Yehui Tang, Chao Zhang, Han Hu, Yunhe Wang
Specifically, we train a tiny student model to match a pre-trained teacher model in the patch-level manifold space.
no code implementations • 29 Jul 2021 • Xianrui Zheng, Chao Zhang, Philip C. Woodland
Furthermore, on the AMI corpus, the proposed conversion for language prior probabilities enables BERT to obtain an extra 3% relative WERR, and the combination of BERT, GPT and GPT-2 results in further improvements.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
1 code implementation • 2 Aug 2021 • Zhen Li, Jing Tang, Deqing Zou, Qian Chen, Shouhuai Xu, Chao Zhang, Yichen Li, Hai Jin
Automatically detecting software vulnerabilities in source code is an important problem that has attracted much attention.
no code implementations • 4 Aug 2021 • Chao Zhang, Sthitie Bom
However, the successful applications of deep learning in soft sensing are still not widely integrated in factory control systems, because most of the research on soft sensing do not have access to large scale industrial data which are varied, noisy and incomplete.
no code implementations • 1 Sep 2021 • Guangzhi Sun, Chao Zhang, Philip C. Woodland
Contextual knowledge is important for real-world automatic speech recognition (ASR) applications.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
no code implementations • 7 Sep 2021 • Sergei Petrov, Chao Zhang, Jaswanth Yella, Yu Huang, Xiaoye Qian, Sthitie Bom
The scope of this challenge is to tackle the task of classifying soft sensing data with machine learning techniques.
no code implementations • 8 Sep 2021 • Yunqi Shao, Florian M. Dietrich, Carl Nettelblad, Chao Zhang
Here we compare the performance of two popular training algorithms, the adaptive moment estimation algorithm (Adam) and the Extended Kalman Filter algorithm (EKF), using the Behler-Parrinello neural network (BPNN) and two publicly accessible datasets of liquid water [Proc.
no code implementations • Findings (EMNLP) 2021 • Yaqing Wang, Haoda Chu, Chao Zhang, Jing Gao
In this work, we study the problem of named entity recognition (NER) in a low resource scenario, focusing on few-shot and zero-shot settings.
no code implementations • 13 Sep 2021 • Chunzhi Gu, Yan Zhao, Chao Zhang
Human motion prediction, which plays a key role in computer vision, generally requires a past motion sequence as input.
no code implementations • Findings (NAACL) 2022 • Simiao Zuo, Yue Yu, Chen Liang, Haoming Jiang, Siawpeng Er, Chao Zhang, Tuo Zhao, Hongyuan Zha
In self-training, the student contributes to the prediction performance, and the teacher controls the training process by generating pseudo-labels.
1 code implementation • 15 Sep 2021 • Harshavardhan Kamarthi, Lingkai Kong, Alexander Rodríguez, Chao Zhang, B. Aditya Prakash
We use CAMul for multiple domains with varied sources and modalities and show that CAMul outperforms other state-of-art probabilistic forecasting models by over 25\% in accuracy and calibration.
no code implementations • 8 Oct 2021 • Zhiyun Lu, Yanwei Pan, Thibault Doutre, Parisa Haghani, Liangliang Cao, Rohit Prabhavalkar, Chao Zhang, Trevor Strohman
Our experiments show that for both losses, the WER on long-form speech reduces substantially as the training utterance length increases.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
1 code implementation • 17 Oct 2021 • Yuefeng Chen, Xiaofeng Mao, Yuan He, Hui Xue, Chao Li, Yinpeng Dong, Qi-An Fu, Xiao Yang, Tianyu Pang, Hang Su, Jun Zhu, Fangcheng Liu, Chao Zhang, Hongyang Zhang, Yichi Zhang, Shilong Liu, Chang Liu, Wenzhao Xiang, Yajie Wang, Huipeng Zhou, Haoran Lyu, Yidan Xu, Zixuan Xu, Taoyu Zhu, Wenjun Li, Xianfeng Gao, Guoqiu Wang, Huanqian Yan, Ying Guo, Chaoning Zhang, Zheng Fang, Yang Wang, Bingyang Fu, Yunfei Zheng, Yekui Wang, Haorong Luo, Zhen Yang
Many works have investigated the adversarial attacks or defenses under the settings where a bounded and imperceptible perturbation can be added to the input.
1 code implementation • 18 Oct 2021 • Yuhui Yuan, Rao Fu, Lang Huang, WeiHong Lin, Chao Zhang, Xilin Chen, Jingdong Wang
We present a High-Resolution Transformer (HRFormer) that learns high-resolution representations for dense prediction tasks, in contrast to the original Vision Transformer that produces low-resolution representations and has high memory and computational cost.
Ranked #3 on Pose Estimation on AIC
no code implementations • 28 Oct 2021 • Chao Zhang, Hanxin Zhang, Atif Khan, Ted Kim, Olasubomi Omoleye, Oluwamayomikun Abiona, Amy Lehman, Christopher O. Olopade, Olufunmilayo I. Olopade, Pedro Lopes, Andrey Rzhetsky
Importance: Lower-resource areas in Africa and Asia face a unique set of healthcare challenges: the dual high burden of communicable and non-communicable diseases; a paucity of highly trained primary healthcare providers in both rural and densely populated urban areas; and a lack of reliable, inexpensive internet connections.
1 code implementation • 2 Nov 2021 • Wenyu Zhu, Zhiyao Feng, Zihan Zhang, Jianjun Chen, Zhijian Ou, Min Yang, Chao Zhang
Recovering binary programs' call graphs is crucial for inter-procedural analysis tasks and applications based on them. transfer One of the core challenges is recognizing targets of indirect calls (i. e., indirect callees).
1 code implementation • 10 Nov 2021 • Chao Zhang, Jaswanth Yella, Yu Huang, Xiaoye Qian, Sergei Petrov, Andrey Rzhetsky, Sthitie Bom
We demonstrate the challenges and effectiveness of modeling industrial big data by a Soft Sensing Transformer model on these data sets.
no code implementations • 12 Nov 2021 • Jaswanth Yella, Chao Zhang, Sergei Petrov, Yu Huang, Xiaoye Qian, Ali A. Minai, Sthitie Bom
Over the last few decades, modern industrial processes have investigated several cost-effective methodologies to improve the productivity and yield of semiconductor manufacturing.
no code implementations • 12 Nov 2021 • Chunzhi Gu, Shuofeng Zhao, Chao Zhang
In this paper, we present a deep generative model based method to generate diverse human motion interpolation results.
no code implementations • 12 Nov 2021 • Yu Huang, Chao Zhang, Jaswanth Yella, Sergei Petrov, Xiaoye Qian, Yufei Tang, Xingquan Zhu, Sthitie Bom
In the era of big data, data-driven based classification has become an essential method in smart manufacturing to guide production and optimize inspection.
no code implementations • 12 Nov 2021 • Weijie Liu, Chao Zhang, Nenggan Zheng, Hui Qian
Optimal transport (OT) naturally arises in a wide range of machine learning applications but may often become the computational bottleneck.
no code implementations • 12 Nov 2021 • Xiaoye Qian, Chao Zhang, Jaswanth Yella, Yu Huang, Ming-Chun Huang, Sthitie Bom
To understand how the proposed model works, the deep visualization approach is applied.
no code implementations • 17 Nov 2021 • Chao Zhang, Zihao Zhang, Mihai Cucuringu, Stefan Zohren
The designed framework circumvents the traditional forecasting step and avoids the estimation of the covariance matrix, lifting the bottleneck for generalizing to a large amount of instruments.
no code implementations • 24 Nov 2021 • Katsuya Hotta, Takuya Akashi, Shogo Tokai, Chao Zhang
Subspace clustering methods which embrace a self-expressive model that represents each data point as a linear combination of other data points in the dataset provide powerful unsupervised learning techniques.
2 code implementations • NeurIPS 2021 • Yuhui Yuan, Rao Fu, Lang Huang, WeiHong Lin, Chao Zhang, Xilin Chen, Jingdong Wang
We present a High-Resolution Transformer (HRFormer) that learns high-resolution representations for dense prediction tasks, in contrast to the original Vision Transformer that produces low-resolution representations and has high memory and computational cost.
no code implementations • 2 Dec 2021 • Chao Zhang, Zhijian Li, Hui Qian, Xin Du
We develop a general Dynamic-weight Particle-based Variational Inference (DPVI) framework according to a novel continuous composite flow, which evolves the positions and weights of particles simultaneously.
1 code implementation • 16 Dec 2021 • Yue Yu, Lingkai Kong, Jieyu Zhang, Rongzhi Zhang, Chao Zhang
We propose {\ours}, a new framework that leverages unlabeled data to improve the label efficiency of active PLM fine-tuning.
no code implementations • 17 Dec 2021 • Yiyuan She, Jiahui Shen, Chao Zhang
In this paper, new information-theoretical limits are presented to reveal the intrinsic cost of seeking for clusters, as well as the blessing from dimensionality in multivariate learning.
no code implementations • 25 Dec 2021 • Rama Cont, Mihai Cucuringu, Chao Zhang
We investigate the impact of order flow imbalance (OFI) on price movements in equity markets in a multi-asset setting.
no code implementations • CVPR 2022 • Jiewen Yang, Xingbo Dong, Liujun Liu, Chao Zhang, Jiajun Shen, Dahai Yu
Besides, the proposed RViT can work on both fixed-length and variant-length video clips properly without requiring large GPU memory thanks to the frame by frame processing flow.
1 code implementation • 4 Jan 2022 • Fangcheng Liu, Chao Zhang, Hongyang Zhang
Extensive experiments verify the effectiveness of our framework on balancing imperceptibility and transferability of the crafted adversarial examples.
no code implementations • 25 Jan 2022 • Chao Zhang, Bo Li, Zhiyun Lu, Tara N. Sainath, Shuo-Yiin Chang
The recurrent neural network transducer (RNN-T) has recently become the mainstream end-to-end approach for streaming automatic speech recognition (ASR).
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
no code implementations • 6 Feb 2022 • Weijie Liu, Chao Zhang, Nenggan Zheng, Hui Qian
In this paper, we propose a novel criterion to measure the graph matching accuracy, structural inconsistency (SI), which is defined based on the network topological structure.
no code implementations • 8 Feb 2022 • Chao Zhang, Yihuang Zhang, Mihai Cucuringu, Zhongmin Qian
We apply machine learning models to forecast intraday realized volatility (RV), by exploiting commonality in intraday volatility via pooling stock data together, and by incorporating a proxy for the market volatility.
1 code implementation • 11 Feb 2022 • Jieyu Zhang, Cheng-Yu Hsieh, Yue Yu, Chao Zhang, Alexander Ratner
Labeling training data has become one of the major roadblocks to using machine learning.
no code implementations • 3 Mar 2022 • Rama Cont, Mihai Cucuringu, Renyuan Xu, Chao Zhang
The estimation of loss distributions for dynamic portfolios requires the simulation of scenarios representing realistic joint dynamics of their components, with particular importance devoted to the simulation of tail risk scenarios.
no code implementations • 7 Mar 2022 • Qi Zhu, Chao Zhang, Chanyoung Park, Carl Yang, Jiawei Han
Then a shift-robust classifier is optimized on training graph and adversarial samples on target graph, which are generated by cluster GNN.
1 code implementation • CVPR 2022 • Xingbo Dong, Wanyan Xu, Zhihui Miao, Lan Ma, Chao Zhang, Jiewen Yang, Zhe Jin, Andrew Beng Jin Teoh, Jiajun Shen
Next, a fully convolutional network is proposed to achieve the low-light image enhancement by fusing colored raw data with synthesized monochrome raw data.
no code implementations • 8 Mar 2022 • Wen Wu, Chao Zhang, Xixin Wu, Philip C. Woodland
In this paper, a novel Bayesian training loss based on per-utterance Dirichlet prior distributions is proposed for verbal emotion recognition, which models the uncertainty in one-hot labels created when human annotators assign the same utterance to different emotion classes.
1 code implementation • 18 Mar 2022 • Rongzhi Zhang, Yue Yu, Pranav Shetty, Le Song, Chao Zhang
Weakly-supervised learning (WSL) has shown promising results in addressing label scarcity on many NLP tasks, but manually designing a comprehensive, high-quality labeling rule set is tedious and difficult.
no code implementations • 18 Mar 2022 • Jun Quan, Ze Wei, Qiang Gan, Jingqi Yao, Jingyi Lu, Yuchen Dong, Yiming Liu, Yi Zeng, Chao Zhang, Yongzhi Li, Huang Hu, Yingying He, Yang Yang, Daxin Jiang
The conversational recommender systems (CRSs) have received extensive attention in recent years.
no code implementations • CVPR 2022 • Yingjie Cai, Kwan-Yee Lin, Chao Zhang, Qiang Wang, Xiaogang Wang, Hongsheng Li
Specifically, we map a series of related partial point clouds into multiple complete shape and occlusion code pairs and fuse the codes to obtain their representations in the unified latent space.
1 code implementation • 30 Mar 2022 • Zhaoyang Huang, Xiaoyu Shi, Chao Zhang, Qiang Wang, Ka Chun Cheung, Hongwei Qin, Jifeng Dai, Hongsheng Li
We introduce optical Flow transFormer, dubbed as FlowFormer, a transformer-based neural network architecture for learning optical flow.
Ranked #1 on Optical Flow Estimation on Sintel-final
no code implementations • NAACL 2022 • Rui Feng, Chen Luo, Qingyu Yin, Bing Yin, Tuo Zhao, Chao Zhang
User sessions empower many search and recommendation tasks on a daily basis.
1 code implementation • 18 May 2022 • Wensheng Li, Chao Zhang, Chuncheng Wang, Hanting Guan, DaCheng Tao
Physics-informed neural networks (PINNs) provide a deep learning framework for numerically solving partial differential equations (PDEs), and have been widely used in a variety of PDE problems.
no code implementations • 18 May 2022 • Guangzhi Sun, Chao Zhang, Philip C Woodland
MBWE and BLMD further improved the effectiveness of TCPGen and achieved more significant WER reductions on the biasing words.
1 code implementation • 27 May 2022 • Yinghao Li, Le Song, Chao Zhang
Weakly supervised named entity recognition methods train label models to aggregate the token annotations of multiple noisy labeling functions (LFs) without seeing any manually annotated labels.
1 code implementation • 16 Jun 2022 • Harshavardhan Kamarthi, Lingkai Kong, Alexander Rodríguez, Chao Zhang, B. Aditya Prakash
We close both these gap and propose PROFHiT, which is a fully probabilistic hierarchical forecasting model that jointly models forecast distribution of entire hierarchy.
no code implementations • 20 Jun 2022 • Guile Wu, Chao Zhang, Stephan Liwicki
In global consistent quantization, we employ contrastive learning for both embedding and quantized representations and fuses these representations for consistent contrastive regularization between instances.
no code implementations • 28 Jun 2022 • Rongzhi Zhang, Rebecca West, Xiquan Cui, Chao Zhang
We develop AMRule, a multi-view rule discovery framework that can (1) adaptively and iteratively discover novel rulers that can complement the current weakly-supervised model to improve compatibility prediction; (2) discover interpretable rules from both structured attribute tables and unstructured product descriptions.
no code implementations • 2 Jul 2022 • Guangzhi Sun, Chao Zhang, Philip C. Woodland
Incorporating biasing words obtained as contextual knowledge is critical for many automatic speech recognition (ASR) applications.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
no code implementations • 4 Jul 2022 • Chunzhi Gu, Jun Yu, Chao Zhang
Specifically, the inductive bias imposed by the extra CVAE path encourages two latent variables in two paths to respectively govern separate representations for each partial-body motion.
no code implementations • 8 Jul 2022 • Xianrui Zheng, Chao Zhang, Philip C. Woodland
Self-supervised-learning-based pre-trained models for speech data, such as Wav2Vec 2. 0 (W2V2), have become the backbone of many speech tasks.
1 code implementation • 16 Jul 2022 • Zizheng Huang, Haoxing Chen, Ziqi Wen, Chao Zhang, Huaxiong Li, Bo wang, Chunlin Chen
Contrastive learning (CL) continuously achieves significant breakthroughs across multiple domains.
8 code implementations • CVPR 2023 • Ding Jia, Yuhui Yuan, Haodi He, Xiaopei Wu, Haojun Yu, WeiHong Lin, Lei Sun, Chao Zhang, Han Hu
One-to-one set matching is a key design for DETR to establish its end-to-end capability, so that object detection does not require a hand-crafted NMS (non-maximum suppression) to remove duplicate detections.
1 code implementation • 7 Aug 2022 • Mengyang Liu, Haozheng Luo, Leonard Thong, Yinghao Li, Chao Zhang, Le Song
Compared to frequently used text annotation tools, our annotation tool allows for the development of weak labels in addition to providing a manual annotation experience.
no code implementations • 29 Aug 2022 • Shuo-Yiin Chang, Bo Li, Tara N. Sainath, Chao Zhang, Trevor Strohman, Qiao Liang, Yanzhang He
This makes doing speech recognition with conversational speech, including one with multiple queries, a challenging task.
no code implementations • 3 Sep 2022 • Chao Zhang, Zijian Tang, Taoming Guo, Jiaxin Lei, Jiaxin Xiao, Anhe Wang, Shuo Bai, Milin Zhang
This paper proposes SaleNet - an end-to-end convolutional neural network (CNN) for sustained attention level evaluation using prefrontal electroencephalogram (EEG).
no code implementations • 13 Sep 2022 • Chao Zhang, Bo Li, Tara Sainath, Trevor Strohman, Sepand Mavandadi, Shuo-Yiin Chang, Parisa Haghani
Language identification is critical for many downstream tasks in automatic speech recognition (ASR), and is beneficial to integrate into multilingual end-to-end ASR as an additional task.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
no code implementations • 15 Sep 2022 • Simiao Zuo, Qingyu Yin, Haoming Jiang, Shaohui Xi, Bing Yin, Chao Zhang, Tuo Zhao
The model subsequently calculates session representations by combining the contextual information with the instant search query using an aggregation network.
1 code implementation • 15 Sep 2022 • Yue Yu, Rongzhi Zhang, ran Xu, Jieyu Zhang, Jiaming Shen, Chao Zhang
Large Language Models have demonstrated remarkable few-shot performance, but the performance can be sensitive to the selection of few-shot instances.
1 code implementation • 27 Sep 2022 • Pranav Shetty, Arunkumar Chitteth Rajan, Christopher Kuenneth, Sonkakshi Gupta, Lakshmi Prerana Panchumarti, Lauren Holm, Chao Zhang, Rampi Ramprasad
The ever-increasing number of materials science articles makes it hard to infer chemistry-structure-property relations from published literature.
no code implementations • 30 Sep 2022 • Hang Zou, Chao Zhang, Samson Lasaulce, Lucas Saludjian, Vincent Poor
The task is modeled by the minimization problem of a general goal function $f(x;g)$ for which the decision $x$ has to be taken from a quantized version of the parameters $g$.
4 code implementations • 3 Oct 2022 • Weicong Liang, Yuhui Yuan, Henghui Ding, Xiao Luo, WeiHong Lin, Ding Jia, Zheng Zhang, Chao Zhang, Han Hu
Vision transformers have recently achieved competitive results across various vision tasks but still suffer from heavy computation costs when processing a large number of tokens.
no code implementations • 9 Oct 2022 • Yukun Zheng, Jiang Bian, Guanghao Meng, Chao Zhang, Honggang Wang, Zhixuan Zhang, Sen Li, Tao Zhuang, Qingwen Liu, Xiaoyi Zeng
These problems promote us to further strengthen the capabilities of our EBR model in both relevance estimation and personalized retrieval.
no code implementations • 26 Oct 2022 • Wei Wang, Chao Zhang, Xiaopei Wu
In this paper, we make use of limited code-switching data as driving materials and explore a shortcut to quickly develop intra-sentential code-switching recognition skill on the commissioned native language acoustic model, where we propose a data-driven method to make the seed lexicon which is used to train grapheme-to-phoneme model to predict mapping pronunciations for foreign language word in code-switching sentences.