6 code implementations • 29 Jul 2019 • Lang Huang, Yuhui Yuan, Jianyuan Guo, Chao Zhang, Xilin Chen, Jingdong Wang
There are two successive attention modules each estimating a sparse affinity matrix.
8 code implementations • 4 Sep 2018 • Yuhui Yuan, Lang Huang, Jianyuan Guo, Chao Zhang, Xilin Chen, Jingdong Wang
To capture richer context information, we further combine our interlaced sparse self-attention scheme with the conventional multi-scale context schemes including pyramid pooling~\citep{zhao2017pyramid} and atrous spatial pyramid pooling~\citep{chen2018deeplab}.
Ranked #9 on Semantic Segmentation on Trans10K
10 code implementations • 12 Nov 2018 • Ling Zhao, Yujiao Song, Chao Zhang, Yu Liu, Pu Wang, Tao Lin, Min Deng, Haifeng Li
However, traffic forecasting has always been considered an open scientific issue, owing to the constraints of urban road network topological structure and the law of dynamic change with time, namely, spatial dependence and temporal dependence.
Ranked #4 on Traffic Prediction on SZ-Taxi
8 code implementations • CVPR 2023 • Ding Jia, Yuhui Yuan, Haodi He, Xiaopei Wu, Haojun Yu, WeiHong Lin, Lei Sun, Chao Zhang, Han Hu
One-to-one set matching is a key design for DETR to establish its end-to-end capability, so that object detection does not require a hand-crafted NMS (non-maximum suppression) to remove duplicate detections.
1 code implementation • 20 Oct 2023 • Changli Tang, Wenyi Yu, Guangzhi Sun, Xianzhao Chen, Tian Tan, Wei Li, Lu Lu, Zejun Ma, Chao Zhang
Hearing is arguably an essential ability of artificial intelligence (AI) agents in the physical world, which refers to the perception and understanding of general auditory information consisting of at least three types of sounds: speech, audio events, and music.
1 code implementation • 3 Mar 2024 • Boyang Wang, Fengyu Yang, Xihang Yu, Chao Zhang, Hanbin Zhao
In addition, we identify two anime-specific challenges of distorted and faint hand-drawn lines and unwanted color artifacts.
1 code implementation • 26 Jan 2024 • Yuhui Li, Fangyun Wei, Chao Zhang, Hongyang Zhang
In this paper, we reconsider speculative sampling and derive two key observations.
1 code implementation • 18 Oct 2021 • Yuhui Yuan, Rao Fu, Lang Huang, WeiHong Lin, Chao Zhang, Xilin Chen, Jingdong Wang
We present a High-Resolution Transformer (HRFormer) that learns high-resolution representations for dense prediction tasks, in contrast to the original Vision Transformer that produces low-resolution representations and has high memory and computational cost.
Ranked #3 on Pose Estimation on AIC
2 code implementations • NeurIPS 2021 • Yuhui Yuan, Rao Fu, Lang Huang, WeiHong Lin, Chao Zhang, Xilin Chen, Jingdong Wang
We present a High-Resolution Transformer (HRFormer) that learns high-resolution representations for dense prediction tasks, in contrast to the original Vision Transformer that produces low-resolution representations and has high memory and computational cost.
1 code implementation • 29 Dec 2023 • Derong Xu, Wei Chen, Wenjun Peng, Chao Zhang, Tong Xu, Xiangyu Zhao, Xian Wu, Yefeng Zheng, Enhong Chen
Information extraction (IE) aims to extract structural knowledge (such as entities, relations, and events) from plain natural language texts.
1 code implementation • 30 Mar 2022 • Zhaoyang Huang, Xiaoyu Shi, Chao Zhang, Qiang Wang, Ka Chun Cheung, Hongwei Qin, Jifeng Dai, Hongsheng Li
We introduce optical Flow transFormer, dubbed as FlowFormer, a transformer-based neural network architecture for learning optical flow.
Ranked #1 on Optical Flow Estimation on Sintel-final
2 code implementations • EMNLP 2020 • Yu Meng, Yunyi Zhang, Jiaxin Huang, Chenyan Xiong, Heng Ji, Chao Zhang, Jiawei Han
In this paper, we explore the potential of only using the label name of each class to train classification models on unlabeled data, without using any labeled documents.
1 code implementation • 28 Jun 2020 • Chen Liang, Yue Yu, Haoming Jiang, Siawpeng Er, Ruijia Wang, Tuo Zhao, Chao Zhang
We study the open-domain named entity recognition (NER) problem under distant supervision.
1 code implementation • 16 May 2020 • Nick Altieri, Rebecca L. Barter, James Duncan, Raaz Dwivedi, Karl Kumbier, Xiao Li, Robert Netzorg, Briton Park, Chandan Singh, Yan Shuo Tan, Tiffany Tang, Yu Wang, Chao Zhang, Bin Yu
We use this data to develop predictions and corresponding prediction intervals for the short-term trajectory of COVID-19 cumulative death counts at the county-level in the United States up to two weeks ahead.
4 code implementations • 26 Apr 2021 • Wei Zeng, Xiaozhe Ren, Teng Su, Hui Wang, Yi Liao, Zhiwei Wang, Xin Jiang, ZhenZhang Yang, Kaisheng Wang, Xiaoda Zhang, Chen Li, Ziyan Gong, Yifan Yao, Xinjing Huang, Jun Wang, Jianfeng Yu, Qi Guo, Yue Yu, Yan Zhang, Jin Wang, Hengtao Tao, Dasen Yan, Zexuan Yi, Fang Peng, Fangqing Jiang, Han Zhang, Lingfeng Deng, Yehong Zhang, Zhe Lin, Chao Zhang, Shaojie Zhang, Mingyue Guo, Shanzhi Gu, Gaojun Fan, YaoWei Wang, Xuefeng Jin, Qun Liu, Yonghong Tian
To enhance the generalization ability of PanGu-$\alpha$, we collect 1. 1TB high-quality Chinese data from a wide range of domains to pretrain the model.
Ranked #1 on Reading Comprehension (One-Shot) on DuReader
Cloze (multi-choices) (Few-Shot) Cloze (multi-choices) (One-Shot) +19
1 code implementation • ECCV 2018 • Yikang Li, Wanli Ouyang, Bolei Zhou, Jianping Shi, Chao Zhang, Xiaogang Wang
Generating scene graph to describe all the relations inside an image gains increasing interests these years.
Ranked #1 on Scene Graph Generation on VRD
1 code implementation • NeurIPS 2023 • Yuchen Zhuang, Yue Yu, Kuan Wang, Haotian Sun, Chao Zhang
To address this issue, we introduce a new dataset called ToolQA, which is designed to faithfully evaluate LLMs' ability to use external tools for question answering.
1 code implementation • NAACL 2021 • Yue Yu, Simiao Zuo, Haoming Jiang, Wendi Ren, Tuo Zhao, Chao Zhang
To address this problem, we develop a contrastive self-training framework, COSINE, to enable fine-tuning LMs with weak supervision.
Ranked #1 on Word Sense Disambiguation on Words in Context
1 code implementation • NeurIPS 2019 • Yu Meng, Jiaxin Huang, Guangyuan Wang, Chao Zhang, Honglei Zhuang, Lance Kaplan, Jiawei Han
While text embeddings are typically learned in the Euclidean space, directional similarity is often more effective in tasks such as word similarity and document clustering, which creates a gap between the training stage and usage stage of text embedding.
1 code implementation • 11 Feb 2022 • Jieyu Zhang, Cheng-Yu Hsieh, Yue Yu, Chao Zhang, Alexander Ratner
Labeling training data has become one of the major roadblocks to using machine learning.
1 code implementation • 2 Jan 2019 • Kai Han, Jianyuan Guo, Chao Zhang, Mingjian Zhu
Based on the considerations above, we propose a novel Attribute-Aware Attention Model ($A^3M$), which can learn local attribute representation and global category representation simultaneously in an end-to-end manner.
Ranked #4 on Fine-Grained Image Classification on CompCars
4 code implementations • NeurIPS 2020 • Lang Huang, Chao Zhang, Hongyang Zhang
We propose self-adaptive training---a new training algorithm that dynamically corrects problematic training labels by model predictions without incurring extra computational cost---to improve generalization of deep learning for potentially corrupted training data.
2 code implementations • 21 Jan 2021 • Lang Huang, Chao Zhang, Hongyang Zhang
We propose self-adaptive training -- a unified training algorithm that dynamically calibrates and enhances training processes by model predictions without incurring an extra computational cost -- to advance both supervised and self-supervised learning of deep neural networks.
1 code implementation • NeurIPS 2023 • Yue Yu, Yuchen Zhuang, Jieyu Zhang, Yu Meng, Alexander Ratner, Ranjay Krishna, Jiaming Shen, Chao Zhang
Large language models (LLMs) have been recently leveraged as training data generators for various natural language processing (NLP) tasks.
1 code implementation • CVPR 2020 • Jianyuan Guo, Kai Han, Yunhe Wang, Chao Zhang, Zhaohui Yang, Han Wu, Xinghao Chen, Chang Xu
To this end, we propose a hierarchical trinity search framework to simultaneously discover efficient architectures for all components (i. e. backbone, neck, and head) of object detector in an end-to-end manner.
2 code implementations • 8 Oct 2019 • Yunqi Shao, Matti Hellström, Pavlin D. Mitev, Lisanne Knijff, Chao Zhang
Atomic neural networks (ANNs) constitute a class of machine learning methods for predicting potential energy surfaces and physico-chemical properties of molecules and materials.
Computational Physics Disordered Systems and Neural Networks Chemical Physics
2 code implementations • ICML 2020 • Lingkai Kong, Jimeng Sun, Chao Zhang
We propose a new method for quantifying uncertainties of DNNs from a dynamical system perspective.
1 code implementation • ICCV 2023 • Xuesong Chen, Shaoshuai Shi, Chao Zhang, Benjin Zhu, Qiang Wang, Ka Chun Cheung, Simon See, Hongsheng Li
3D multi-object tracking (MOT) is vital for many applications including autonomous driving vehicles and service robots.
1 code implementation • ICCV 2017 • Yuhui Yuan, Kuiyuan Yang, Chao Zhang
This motivates us to ensemble a set of models with different complexities in cascaded manner and mine hard examples adaptively, a sample is judged by a series of models with increasing complexities and only updates models that consider the sample as a hard case.
Ranked #14 on Image Retrieval on SOP
1 code implementation • 14 Jul 2023 • XueMei Dong, Chao Zhang, Yuhang Ge, YUREN MAO, Yunjun Gao, Lu Chen, Jinshu Lin, Dongfang Lou
This paper proposes a ChatGPT-based zero-shot Text-to-SQL method, dubbed C3, which achieves 82. 3\% in terms of execution accuracy on the holdout test set of Spider and becomes the state-of-the-art zero-shot Text-to-SQL method on the Spider Challenge.
Ranked #4 on Text-To-SQL on spider
1 code implementation • 29 Dec 2018 • Yu Meng, Jiaming Shen, Chao Zhang, Jiawei Han
During the training process, our model features a hierarchical neural structure, which mimics the given hierarchy and is capable of determining the proper levels for documents with a blocking mechanism.
1 code implementation • CVPR 2022 • Xingbo Dong, Wanyan Xu, Zhihui Miao, Lan Ma, Chao Zhang, Jiewen Yang, Zhe Jin, Andrew Beng Jin Teoh, Jiajun Shen
Next, a fully convolutional network is proposed to achieve the low-light image enhancement by fusing colored raw data with synthesized monochrome raw data.
1 code implementation • CVPR 2019 • Chao Zhang, Shuaicheng Liu, Xun Xu, Ce Zhu
Recently, MobileNets and ShuffleNets have been proposed to reduce the number of parameters, yielding lightweight models.
Ranked #3 on Age Estimation on FGNET
1 code implementation • CVPR 2017 • Chao Zhang, Sergi Pujades, Michael Black, Gerard Pons-Moll
We address the problem of estimating human pose and body shape from 3D scans over time.
1 code implementation • 19 Jan 2024 • Yuchen Hu, Chen Chen, Chao-Han Huck Yang, Ruizhe Li, Chao Zhang, Pin-Yu Chen, EnSiong Chng
To this end, we propose to extract a language-space noise embedding from the N-best list to represent the noise conditions of source speech, which can promote the denoising process in GER.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +6
1 code implementation • 22 Oct 2019 • Qiujia Li, Florian L. Kreyssig, Chao Zhang, Philip C. Woodland
In this paper, we propose Discriminative Neural Clustering (DNC) that formulates data clustering with a maximum number of clusters as a supervised sequence-to-sequence learning problem.
1 code implementation • 4 Oct 2020 • Yue Yu, Kexin Huang, Chao Zhang, Lucas M. Glass, Jimeng Sun, Cao Xiao
Furthermore, most previous works focus on binary DDI prediction whereas the multi-typed DDI pharmacological effect prediction is a more meaningful but harder task.
1 code implementation • 2 Sep 2018 • Yu Meng, Jiaming Shen, Chao Zhang, Jiawei Han
Although many semi-supervised and weakly-supervised text classification models exist, they cannot be easily applied to deep neural models and meanwhile support limited supervision types.
1 code implementation • NeurIPS 2023 • Haotian Sun, Yuchen Zhuang, Lingkai Kong, Bo Dai, Chao Zhang
We propose a closed-loop approach, AdaPlanner, which allows the LLM agent to refine its self-generated plan adaptively in response to environmental feedback.
1 code implementation • 15 Jun 2023 • Ziyang Ma, Zhisheng Zheng, Guanrou Yang, Yu Wang, Chao Zhang, Xie Chen
Our models outperform other SSL models significantly on the LibriSpeech benchmark without the need for iterative re-clustering and re-training.
1 code implementation • 10 Jul 2018 • Yu Shi, Qi Zhu, Fang Guo, Chao Zhang, Jiawei Han
To cope with the challenges in the comprehensive transcription of HINs, we propose the HEER algorithm, which embeds HINs via edge representations that are further coupled with properly-learned heterogeneous metrics.
1 code implementation • ICCV 2019 • Jianyuan Guo, Yuhui Yuan, Lang Huang, Chao Zhang, Jinge Yao, Kai Han
On the other hand, there still exist many useful contextual cues that do not fall into the scope of predefined human parts or attributes.
Ranked #59 on Person Re-Identification on DukeMTMC-reID
1 code implementation • 13 Sep 2023 • Yuhui Li, Fangyun Wei, Jinjing Zhao, Chao Zhang, Hongyang Zhang
We discover that by integrating self-evaluation and rewind mechanisms, unaligned LLMs can directly produce responses consistent with human preferences via self-boosting.
1 code implementation • 18 Jul 2020 • Yu Meng, Yunyi Zhang, Jiaxin Huang, Yu Zhang, Chao Zhang, Jiawei Han
Mining a set of meaningful topics organized into a hierarchy is intuitively appealing since topic correlations are ubiquitous in massive text corpora.
Ranked #1 on Topic Models on Arxiv HEP-TH citation graph
1 code implementation • 2 Jun 2023 • Guangzhi Sun, Xianrui Zheng, Chao Zhang, Philip C. Woodland
End-to-end automatic speech recognition (ASR) and large language models, such as Whisper and GPT-2, have recently been scaled to use vast amounts of training data.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
1 code implementation • 20 Aug 2019 • Yu Meng, Jiaxin Huang, Guangyuan Wang, Zihan Wang, Chao Zhang, Yu Zhang, Jiawei Han
We propose a new task, discriminative topic mining, which leverages a set of user-provided category names to mine discriminative topics from text corpora.
2 code implementations • NeurIPS 2018 • Shupeng Su, Chao Zhang, Kai Han, Yonghong Tian
To convert the input into binary code, hashing algorithm has been widely used for approximate nearest neighbor search on large-scale image sets due to its computation and storage efficiency.
1 code implementation • 2 Mar 2017 • Felipe Petroski Such, Shagan Sah, Miguel Dominguez, Suhas Pillai, Chao Zhang, Andrew Michael, Nathan Cahill, Raymond Ptucha
Graph-CNNs can handle both heterogeneous and homogeneous graph data, including graphs having entirely different vertex or edge sets.
1 code implementation • 25 Nov 2020 • Wei Wang, Chao Zhang, Xiaopei Wu
Accent recognition with deep learning framework is a similar work to deep speaker identification, they're both expected to give the input speech an identifiable representation.
1 code implementation • 27 Oct 2022 • Yue Yu, Chenyan Xiong, Si Sun, Chao Zhang, Arnold Overwijk
We present a new zero-shot dense retrieval (ZeroDR) method, COCO-DR, to improve the generalization ability of dense retrieval by combating the distribution shifts between source training tasks and target scenarios.
Ranked #1 on Zero-shot Text Search on CQADupStack
1 code implementation • EMNLP 2020 • Rongzhi Zhang, Yue Yu, Chao Zhang
Our method, SeqMix, simply augments the queried samples by generating extra labeled sequences in each iteration.
1 code implementation • 21 Nov 2023 • Zeyu Gao, Hao Wang, Yuchen Zhou, Wenyu Zhu, Chao Zhang
Given the significant successes of large language models (LLMs) in various tasks, there is growing anticipation of their efficacy in vulnerability detection.
1 code implementation • EMNLP 2020 • Lingkai Kong, Haoming Jiang, Yuchen Zhuang, Jie Lyu, Tuo Zhao, Chao Zhang
Fine-tuned pre-trained language models can suffer from severe miscalibration for both in-distribution and out-of-distribution (OOD) data due to over-parameterization.
1 code implementation • 18 Mar 2022 • Rongzhi Zhang, Yue Yu, Pranav Shetty, Le Song, Chao Zhang
Weakly-supervised learning (WSL) has shown promising results in addressing label scarcity on many NLP tasks, but manually designing a comprehensive, high-quality labeling rule set is tedious and difficult.
1 code implementation • ACL 2022 • Rongzhi Zhang, Yue Yu, Pranav Shetty, Le Song, Chao Zhang
Weakly-supervised learning (WSL) has shown promising results in addressing label scarcity on many NLP tasks, but manually designing a comprehensive, high-quality labeling rule set is tedious and difficult.
1 code implementation • 2 Nov 2021 • Wenyu Zhu, Zhiyao Feng, Zihan Zhang, Jianjun Chen, Zhijian Ou, Min Yang, Chao Zhang
Recovering binary programs' call graphs is crucial for inter-procedural analysis tasks and applications based on them. transfer One of the core challenges is recognizing targets of indirect calls (i. e., indirect callees).
2 code implementations • ACL 2021 • Yinghao Li, Pranav Shetty, Lucas Liu, Chao Zhang, Le Song
To address this challenge, we propose a conditional hidden Markov model (CHMM), which can effectively infer true labels from multi-source noisy labels in an unsupervised way.
1 code implementation • 16 Jun 2022 • Harshavardhan Kamarthi, Lingkai Kong, Alexander Rodríguez, Chao Zhang, B. Aditya Prakash
We close both these gap and propose PROFHiT, which is a fully probabilistic hierarchical forecasting model that jointly models forecast distribution of entire hierarchy.
1 code implementation • 17 Oct 2023 • Harshavardhan Kamarthi, Lingkai Kong, Alexander Rodríguez, Chao Zhang, B. Aditya Prakash
We close both these gap and propose PROFHiT, which is a fully probabilistic hierarchical forecasting model that jointly models forecast distribution of entire hierarchy.
1 code implementation • ICCV 2019 • Chao Zhang, Stephan Liwicki, William Smith, Roberto Cipolla
For the spherical domain, several methods recently adopt an icosahedron mesh, but systems are typically rotation invariant or require significant memory and parameters, thus enabling execution only at very low resolutions.
Ranked #21 on Semantic Segmentation on Stanford2D3D Panoramic
1 code implementation • NeurIPS 2021 • Qi Zhu, Carl Yang, Yidan Xu, Haonan Wang, Chao Zhang, Jiawei Han
Graph neural networks (GNNs) have achieved superior performance in various applications, but training dedicated GNNs can be costly for large-scale graphs.
2 code implementations • 22 Dec 2018 • Chao Zhang, Fangbo Tao, Xiusi Chen, Jiaming Shen, Meng Jiang, Brian Sadler, Michelle Vanni, Jiawei Han
Our method, TaxoGen, uses term embeddings and hierarchical clustering to construct a topic taxonomy in a recursive fashion.
Databases
1 code implementation • 17 Oct 2021 • Yuefeng Chen, Xiaofeng Mao, Yuan He, Hui Xue, Chao Li, Yinpeng Dong, Qi-An Fu, Xiao Yang, Tianyu Pang, Hang Su, Jun Zhu, Fangcheng Liu, Chao Zhang, Hongyang Zhang, Yichi Zhang, Shilong Liu, Chang Liu, Wenzhao Xiang, Yajie Wang, Huipeng Zhou, Haoran Lyu, Yidan Xu, Zixuan Xu, Taoyu Zhu, Wenjun Li, Xianfeng Gao, Guoqiu Wang, Huanqian Yan, Ying Guo, Chaoning Zhang, Zheng Fang, Yang Wang, Bingyang Fu, Yunfei Zheng, Yekui Wang, Haorong Luo, Zhen Yang
Many works have investigated the adversarial attacks or defenses under the settings where a bounded and imperceptible perturbation can be added to the input.
1 code implementation • 4 Jan 2022 • Fangcheng Liu, Chao Zhang, Hongyang Zhang
Extensive experiments verify the effectiveness of our framework on balancing imperceptibility and transferability of the crafted adversarial examples.
1 code implementation • 3 Jul 2021 • Zhiwei Hao, Jianyuan Guo, Ding Jia, Kai Han, Yehui Tang, Chao Zhang, Han Hu, Yunhe Wang
Specifically, we train a tiny student model to match a pre-trained teacher model in the patch-level manifold space.
1 code implementation • 26 Feb 2024 • Hao Wang, Zeyu Gao, Chao Zhang, Zihan Sha, Mingyang Sun, Yuchen Zhou, Wenyu Zhu, Wenju Sun, Han Qiu, Xi Xiao
At the core, our approach boosts superior transfer learning capabilities by effectively aligning binary code with their semantics explanations (in natural language), resulting a model able to generate better embeddings for binary code.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Wendi Ren, Yinghao Li, Hanting Su, David Kartchner, Cassie Mitchell, Chao Zhang
We study the problem of learning neural text classifiers without using any labeled data, but only easy-to-provide rules as multiple weak supervision sources.
1 code implementation • 27 Sep 2022 • Pranav Shetty, Arunkumar Chitteth Rajan, Christopher Kuenneth, Sonkakshi Gupta, Lakshmi Prerana Panchumarti, Lauren Holm, Chao Zhang, Rampi Ramprasad
The ever-increasing number of materials science articles makes it hard to infer chemistry-structure-property relations from published literature.
1 code implementation • 9 Nov 2022 • Wen Wu, Chao Zhang, Philip C. Woodland
Automatic emotion recognition in conversation (ERC) is crucial for emotion-aware conversational artificial intelligence.
1 code implementation • 6 Jan 2024 • Zeju Li, Chao Zhang, Xiaoyan Wang, Ruilong Ren, Yifan Xu, Ruifei Ma, Xiangde Liu
The remarkable potential of multi-modal large language models (MLLMs) in comprehending both vision and language information has been widely acknowledged.
1 code implementation • 15 Sep 2022 • Yue Yu, Rongzhi Zhang, ran Xu, Jieyu Zhang, Jiaming Shen, Chao Zhang
Large Language Models have demonstrated remarkable few-shot performance, but the performance can be sensitive to the selection of few-shot instances.
1 code implementation • 18 May 2023 • Yue Yu, Yuchen Zhuang, Rongzhi Zhang, Yu Meng, Jiaming Shen, Chao Zhang
With the development of large language models (LLMs), zero-shot learning has attracted much attention for various NLP tasks.
Ranked #1 on Zero-Shot Text Classification on AG News
1 code implementation • 18 Jun 2020 • Yue Yu, Yinghao Li, Jiaming Shen, Hao Feng, Jimeng Sun, Chao Zhang
We propose a self-supervised taxonomy expansion model named STEAM, which leverages natural supervision in the existing taxonomy for expansion.
1 code implementation • 24 Aug 2023 • Wenyu Zhu, Hao Wang, Yuchen Zhou, JiaMing Wang, Zihan Sha, Zeyu Gao, Chao Zhang
By feeding explicit knowledge as additional inputs to the Transformer, and fusing implicit knowledge with a novel pre-training task, kTrans provides a new perspective to incorporating domain knowledge into a Transformer framework.
1 code implementation • 1 Jul 2021 • Qiujia Li, Chao Zhang, Philip C. Woodland
Commonly used automatic speech recognition (ASR) systems can be classified into frame-synchronous and label-synchronous categories, based on whether the speech is decoded on a per-frame or per-label basis.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
1 code implementation • 23 Oct 2017 • Kai Han, Yunhe Wang, Chao Zhang, Chao Li, Chao Xu
High-dimensional data in many areas such as computer vision and machine learning tasks brings in computational and analytical difficulty.
1 code implementation • 25 Nov 2022 • Lingkai Kong, Jiaming Cui, Yuchen Zhuang, Rui Feng, B. Aditya Prakash, Chao Zhang
Decision-focused learning (DFL) was recently proposed for stochastic optimization problems that involve unknown parameters.
1 code implementation • ICCV 2023 • Jianshuo Dong, Han Qiu, Yiming Li, Tianwei Zhang, Yuanjie Li, Zeqi Lai, Chao Zhang, Shu-Tao Xia
We propose a training-assisted bit flip attack, in which the adversary is involved in the training stage to build a high-risk model to release.
1 code implementation • 15 Sep 2021 • Harshavardhan Kamarthi, Lingkai Kong, Alexander Rodríguez, Chao Zhang, B. Aditya Prakash
We use CAMul for multiple domains with varied sources and modalities and show that CAMul outperforms other state-of-art probabilistic forecasting models by over 25\% in accuracy and calibration.
1 code implementation • 16 Dec 2021 • Yue Yu, Lingkai Kong, Jieyu Zhang, Rongzhi Zhang, Chao Zhang
We propose {\ours}, a new framework that leverages unlabeled data to improve the label efficiency of active PLM fine-tuning.
1 code implementation • NAACL 2022 • Yue Yu, Lingkai Kong, Jieyu Zhang, Rongzhi Zhang, Chao Zhang
We develop AcTune, a new framework that improves the label efficiency of active PLM fine-tuning by unleashing the power of unlabeled data via self-training.
1 code implementation • 5 Oct 2020 • Yinghao Li, Rui Feng, Isaac Rehg, Chao Zhang
We study the problem of using (partial) constituency parse trees as syntactic guidance for controlled text generation.
1 code implementation • CVPR 2021 • Yingjie Cai, Xuesong Chen, Chao Zhang, Kwan-Yee Lin, Xiaogang Wang, Hongsheng Li
The key insight is that we decouple the instances from a coarsely completed semantic scene instead of a raw input image to guide the reconstruction of instances and the overall scene.
Ranked #1 on 3D Semantic Scene Completion on NYUv2
1 code implementation • 10 Jan 2023 • ran Xu, Yue Yu, Hejie Cui, Xuan Kan, Yanqiao Zhu, Joyce Ho, Chao Zhang, Carl Yang
Our further analysis demonstrates that our proposed data selection strategy reduces the noise of pseudo labels by 36. 8% and saves 57. 3% of the time when compared with the best baseline.
1 code implementation • 28 Oct 2022 • Zihan Zhang, Jinfeng Li, Ning Shi, Bo Yuan, Xiangyu Liu, Rong Zhang, Hui Xue, Donghong Sun, Chao Zhang
Despite of the superb performance on a wide range of tasks, pre-trained language models (e. g., BERT) have been proved vulnerable to adversarial texts.
1 code implementation • NeurIPS 2021 • Harshavardhan Kamarthi, Lingkai Kong, Alexander Rodríguez, Chao Zhang, B. Aditya Prakash
We model the forecasting task as a probabilistic generative process and propose a functional neural process model called EPIFNP, which directly models the probability density of the forecast value.
4 code implementations • 3 Oct 2022 • Weicong Liang, Yuhui Yuan, Henghui Ding, Xiao Luo, WeiHong Lin, Ding Jia, Zheng Zhang, Chao Zhang, Han Hu
Vision transformers have recently achieved competitive results across various vision tasks but still suffer from heavy computation costs when processing a large number of tokens.
2 code implementations • 9 Oct 2023 • Guangzhi Sun, Wenyi Yu, Changli Tang, Xianzhao Chen, Tian Tan, Wei Li, Lu Lu, Zejun Ma, Chao Zhang
Audio-visual large language models (LLM) have drawn significant attention, yet the fine-grained combination of both input streams is rather under-explored, which is challenging but necessary for LLMs to understand general video inputs.
1 code implementation • 13 Nov 2023 • Jerry Junyang Cheung, Yuchen Zhuang, Yinghao Li, Pranav Shetty, Wantian Zhao, Sanjeev Grampurohit, Rampi Ramprasad, Chao Zhang
Scientific information extraction (SciIE), which aims to automatically extract information from scientific literature, is becoming more important than ever.
1 code implementation • 5 Oct 2020 • Wanzheng Zhu, Chao Zhang, Shuochao Yao, Xiaobin Gao, Jiawei Han
We propose SHMM, a multi-modal spherical hidden Markov model for semantics-rich human mobility modeling.
1 code implementation • 30 May 2023 • Yuchen Zhuang, Yue Yu, Lingkai Kong, Xiang Chen, Chao Zhang
Most existing methods for learning from noisy labels use static input features for denoising, but these methods are limited by the information they can provide on true label distributions and can result in biased or incorrect predictions.
1 code implementation • 11 Jun 2023 • Wen Wu, Chao Zhang, Philip C. Woodland
In automatic emotion recognition (AER), labels assigned by different human annotators to the same utterance are often inconsistent due to the inherent complexity of emotion and the subjectivity of perception.
1 code implementation • 17 Jul 2023 • Lingkai Kong, Jiaming Cui, Haotian Sun, Yuchen Zhuang, B. Aditya Prakash, Chao Zhang
However, existing diffusion-based graph generative models are mostly one-shot generative models that apply Gaussian diffusion in the dequantized adjacency matrix space.
1 code implementation • 13 Feb 2024 • Haotian Sun, Yuchen Zhuang, Wei Wei, Chao Zhang, Bo Dai
BBox-Adapter distinguishes target and source domain data by treating target data as positive and source data as negative.
1 code implementation • 27 May 2022 • Yinghao Li, Le Song, Chao Zhang
Weakly supervised named entity recognition methods train label models to aggregate the token annotations of multiple noisy labeling functions (LFs) without seeing any manually annotated labels.
1 code implementation • 1 Nov 2022 • Yue Yu, Xuan Kan, Hejie Cui, ran Xu, Yujia Zheng, Xiangchen Song, Yanqiao Zhu, Kun Zhang, Razieh Nabi, Ying Guo, Chao Zhang, Carl Yang
To better adapt GNNs for fMRI analysis, we propose TBDS, an end-to-end framework based on \underline{T}ask-aware \underline{B}rain connectivity \underline{D}AG (short for Directed Acyclic Graph) \underline{S}tructure generation for fMRI analysis.
1 code implementation • 14 Aug 2023 • Wen Wu, Chao Zhang, Philip C. Woodland
Two metrics are proposed to evaluate AER performance with automatic segmentation based on time-weighted emotion and speaker classification errors.
1 code implementation • 6 Oct 2023 • Wei Lv, Chao Zhang, Huaxiong Li, Xiuyi Jia, Chunlin Chen
We further consider the graph noise of projected data caused by missing samples and use a tensor-decomposition based graph filter for robust clustering. JPLTD decomposes the original tensor into an intrinsic tensor and a sparse tensor.
1 code implementation • 16 Jul 2022 • Zizheng Huang, Haoxing Chen, Ziqi Wen, Chao Zhang, Huaxiong Li, Bo wang, Chunlin Chen
Contrastive learning (CL) continuously achieves significant breakthroughs across multiple domains.
1 code implementation • 29 Oct 2022 • Guangzhi Sun, Chao Zhang, Philip C. Woodland
Specifically, a tree-constrained pointer generator (TCPGen), a powerful and efficient biasing model component, is studied, which leverages a slot shortlist with corresponding entities to extract biasing lists.
1 code implementation • 30 May 2023 • Guangzhi Sun, Chao Zhang, Phil Woodland
The incorporation of biasing words obtained through contextual knowledge is of paramount importance in automatic speech recognition (ASR) applications.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
2 code implementations • 14 Jun 2023 • Yinghao Li, Lingkai Kong, Yuanqi Du, Yue Yu, Yuchen Zhuang, Wenhao Mu, Chao Zhang
While some studies have included UQ to improve molecular pre-trained models, the process of selecting suitable backbone and UQ methods for reliable molecular uncertainty estimation remains underexplored.
1 code implementation • 17 Mar 2024 • Yuzhao Heng, Chunyuan Deng, Yitong Li, Yue Yu, Yinghao Li, Rongzhi Zhang, Chao Zhang
Although Large Language Models (LLMs) exhibit remarkable adaptability across domains, these models often fall short in structured knowledge extraction tasks such as named entity recognition (NER).
1 code implementation • 2 Aug 2021 • Zhen Li, Jing Tang, Deqing Zou, Qian Chen, Shouhuai Xu, Chao Zhang, Yichen Li, Hai Jin
Automatically detecting software vulnerabilities in source code is an important problem that has attracted much attention.
1 code implementation • 26 Oct 2022 • Yuchen Zhuang, Yinghao Li, Jerry Junyang Cheung, Yue Yu, Yingjun Mou, Xiang Chen, Le Song, Chao Zhang
We study the problem of extracting N-ary relation tuples from scientific articles.
1 code implementation • 10 Oct 2019 • Wanzheng Zhu, Hongyu Gong, Jiaming Shen, Chao Zhang, Jingbo Shang, Suma Bhat, Jiawei Han
In this paper, we study the task of multi-faceted set expansion, which aims to capture all semantic facets in the seed set and return multiple sets of entities, one for each semantic facet.
1 code implementation • 10 Nov 2021 • Chao Zhang, Jaswanth Yella, Yu Huang, Xiaoye Qian, Sergei Petrov, Andrey Rzhetsky, Sthitie Bom
We demonstrate the challenges and effectiveness of modeling industrial big data by a Soft Sensing Transformer model on these data sets.
1 code implementation • 17 Sep 2023 • Qiuming Zhao, Guangzhi Sun, Chao Zhang, Mingxing Xu, Thomas Fang Zheng
Recent end-to-end automatic speech recognition (ASR) models have become increasingly larger, making them particularly challenging to be deployed on resource-constrained devices.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
1 code implementation • 13 Nov 2023 • Yinghao Li, Haorui Wang, Chao Zhang
Large Language Models (LLMs) have shown remarkable proficiency in language understanding and have been successfully applied to a variety of real-world tasks through task-specific fine-tuning or prompt engineering.
1 code implementation • 20 Feb 2024 • Yinghao Li, Rampi Ramprasad, Chao Zhang
It breaks the generation into a two-step pipeline: initially, LLMs generate answers in natural language as intermediate responses.
1 code implementation • 26 Nov 2022 • Yuhui Li, Zejia Wu, Chao Zhang, Hongyang Zhang
In this work, we introduce the concepts of direct and indirect effects from causal inference to the domain generalization problem.
1 code implementation • 29 May 2021 • Jiahao Xie, Chao Zhang, Zebang Shen, Weijie Liu, Hui Qian
We establish theoretical guarantees of CDMA under different choices of hyperparameters and conduct experiments on AUC maximization, robust adversarial network training, and GAN training tasks.
1 code implementation • 30 Sep 2023 • Wen Wu, Wenlin Chen, Chao Zhang, Philip C. Woodland
Human annotator simulation (HAS) serves as a cost-effective substitute for human evaluation such as data annotation and system assessment.
1 code implementation • 29 Feb 2024 • Pranav Shetty, Aishat Adeboye, Sonakshi Gupta, Chao Zhang, Rampi Ramprasad
We present a natural language processing pipeline that was used to extract polymer solar cell property data from the literature and simulate various active learning strategies.
1 code implementation • 26 Jun 2023 • Chao Zhang, Shiwei Wu, Sirui Zhao, Tong Xu, Enhong Chen
In this paper, we present a solution for enhancing video alignment to improve multi-step inference.
1 code implementation • 25 Oct 2023 • Zichong Li, Yanbo Xu, Simiao Zuo, Haoming Jiang, Chao Zhang, Tuo Zhao, Hongyuan Zha
We conduct extensive experiments in both event type prediction and uncertainty quantification of arrival time.
no code implementations • 18 Jun 2018 • Chao Zhang, Philip Woodland
Gating is a key technique used for integrating information from multiple sources by long short-term memory (LSTM) models and has recently also been applied to other models such as the highway network.
no code implementations • 8 May 2018 • Chao Zhang, Ce Zhu, Jimin Xiao, Xun Xu, Yipeng Liu
Finally we demonstrate the effectiveness of both approaches by visualizing the Class Activation Map (CAM) and discover that grid dropout is more aware of the whole facial areas and more robust than neuron dropout for small training dataset.
no code implementations • 22 Oct 2016 • Julie Yixuan Zhu, Chao Zhang, Huichu Zhang, Shi Zhi, Victor O. K. Li, Jiawei Han, Yu Zheng
Therefore, we present \emph{p-Causality}, a novel pattern-aided causality analysis approach that combines the strengths of \emph{pattern mining} and \emph{Bayesian learning} to efficiently and faithfully identify the \emph{ST causal pathways}.
no code implementations • 22 Feb 2018 • Chao Zhang, Philip Woodland
Vanishing long-term gradients are a major issue in training standard recurrent neural networks (RNNs), which can be alleviated by long short-term memory (LSTM) models with memory cells.
no code implementations • 18 Feb 2018 • Florian Kreyssig, Chao Zhang, Philip Woodland
Time delay neural networks (TDNNs) are an effective acoustic model for large vocabulary speech recognition.
no code implementations • ICLR 2019 • Yihan Gao, Chao Zhang, Jian Peng, Aditya Parameswaran
Both theoretical and empirical evidence are provided to support this argument: (a) we prove that the generalization error of these methods can be bounded by limiting the norm of vectors, regardless of the embedding dimension; (b) we show that the generalization performance of linear graph embedding methods is correlated with the norm of embedding vectors, which is small due to the early stopping of SGD and the vanishing gradients.
no code implementations • 22 Jan 2018 • Sibo Yang, Chao Zhang, Wei Wu
Considered in this short note is the design of output layer nodes of feedforward neural networks for solving multi-class classification problems with r (bigger than or equal to 3) classes of samples.
no code implementations • 9 Sep 2017 • Shuochao Yao, Yiran Zhao, Huajie Shao, Aston Zhang, Chao Zhang, Shen Li, Tarek Abdelzaher
Recent advances in deep learning have led various applications to unprecedented achievements, which could potentially bring higher intelligence to a broad spectrum of mobile and ubiquitous applications.
no code implementations • ICLR 2018 • Yuhui Yuan, Kuiyuan Yang, Chao Zhang
Thus, we propose feature incay to also regularize representation learning, which favors feature vectors with large norm when the samples can be correctly classified.
no code implementations • 21 Nov 2016 • Nathan D. Cahill, Harmeet Singh, Chao Zhang, Daryl A. Corcoran, Alison M. Prengaman, Paul S. Wenger, John F. Hamilton, Peter Bajorski, Andrew M. Michael
Functional connectivity analysis yields powerful insights into our understanding of the human brain.
no code implementations • 13 Nov 2016 • Zebang Shen, Hui Qian, Chao Zhang, Tengfei Zhou
Algorithms with fast convergence, small number of data access, and low per-iteration complexity are particularly favorable in the big data era, due to the demand for obtaining \emph{highly accurate solutions} to problems with \emph{a large number of samples} in \emph{ultra-high} dimensional space.
no code implementations • 25 May 2016 • Yanxiang Chen, Yuxing Hu, Luming Zhang, Ping Li, Chao Zhang
To remedy these problems, we develop a deep architecture to learn aesthetically-relevant visual attributes from Flickr1, which are localized by multiple textual attributes in a weakly-supervised setting.
no code implementations • 25 Jun 2015 • Hongyang Zhang, Zhouchen Lin, Chao Zhang
As an application, we also find that the solutions to extended robust Low-Rank Representation and to our extended robust MC are mutually expressible, so both our theory and algorithm can be applied to the subspace clustering problem with missing values under certain conditions.
1 code implementation • 4 Jan 2015 • Chao Zhang, Chunhua Shen, Tingzhi Shen
We experimentally demonstrate that the learned features, together with our matching model, outperforms state-of-the-art methods such as the SIFT flow, coherency sensitive hashing and the recent deformable spatial pyramid matching methods both in terms of accuracy and computation efficiency.
no code implementations • 6 Dec 2014 • Hongyang Zhang, Zhouchen Lin, Chao Zhang, Junbin Gao
More specifically, we discover that once a solution to one of the models is obtained, we can obtain the solutions to other models in closed-form formulations.
no code implementations • 28 Aug 2014 • Chao Zhang, DaCheng Tao, Tao Hu, Xiang Li
We are mainly concerned with two theoretical questions: 1) under what conditions does RMTL perform better with a smaller task sample size than STL?
no code implementations • 19 May 2014 • Chao Zhang, Hong-cen Mei, Hao Yang
A large number of experimental data shows that Support Vector Machine (SVM) algorithm has obvious advantages in text classification, handwriting recognition, image classification, bioinformatics, and some other fields.
no code implementations • 2 Jan 2014 • Chao Zhang, Lei Zhang, Wei Fan, Jieping Ye
Finally, we analyze the asymptotic convergence and the rate of convergence of the learning process for representative domain adaptation.
no code implementations • 26 Sep 2013 • Chao Zhang
samples, and then show that the generalization bounds have a faster rate of convergence than the traditional results.
no code implementations • 23 Apr 2013 • Hongyang Zhang, Zhouchen Lin, Chao Zhang
For several rank minimization problems, such a replacement has been theoretically proven to be valid, i. e., the solution to nuclear norm minimization problem is also the solution to rank minimization problem.
no code implementations • NeurIPS 2012 • Chao Zhang, Lei Zhang, Jieping Ye
Afterwards, we analyze the asymptotic convergence and the rate of convergence of the learning process for such kind of domain adaptation.
no code implementations • 5 Sep 2018 • Chao Zhang, Xuequan Lu, Takuya Akashi
To settle this issue, we propose a blur-countering method for detecting valid keypoints for various types and degrees of blurred images.
no code implementations • NeurIPS 2018 • Zhisheng Zhong, Tiancheng Shen, Yibo Yang, Zhouchen Lin, Chao Zhang
To solve these problems, we propose the Super-Resolution CliqueNet (SRCliqueNet) to reconstruct the high resolution (HR) image with better textural details in the wavelet domain.
no code implementations • NeurIPS 2018 • Yiwen Guo, Chao Zhang, Chang-Shui Zhang, Yurong Chen
Deep neural networks (DNNs) are computationally/memory-intensive and vulnerable to adversarial attacks, making them prohibitive in some real-world applications.
no code implementations • 11 Nov 2018 • Chao Zhang, Min-Hsiu Hsieh, DaCheng Tao
First, we prove that the complexity of function classes convolving with vicinal functions can be controlled by that of the original function classes under the assumption that the function class is composed of Lipschitz-continuous functions.
no code implementations • 26 Nov 2018 • Bold Naranchimeg, Chao Zhang, Takuya Akashi
In recent decade, many state-of-the-art algorithms on image classification as well as audio classification have achieved noticeable successes with the development of deep convolutional neural network (CNN).
Audio Classification Bird Species Classification With Audio-Visual Data +3
no code implementations • ICLR 2019 • Yaohua Tang, Kaixiang Mo, Qian Xu, Chao Zhang, Qiang Yang
When building models for novel natural language domains, a major challenge is the lack of data in the new domains, no matter whether the data is annotated or not.
no code implementations • CVPR 2014 • Guodong Guo, Chao Zhang
Further, we study the amount of data needed in the target population to learn a cross-population age estimator.
no code implementations • CVPR 2015 • Junchi Yan, Chao Zhang, Hongyuan Zha, Wei Liu, Xiaokang Yang, Stephen M. Chu
Evaluations on both synthetic and real-world data corroborate the efficiency of our method.
no code implementations • CVPR 2016 • Chao Zhang, William A. P. Smith, Arnaud Dessein, Nick Pears, Hang Dai
In this paper we present a method for computing dense correspondence between a set of 3D face meshes using functional maps.
no code implementations • ICCV 2015 • Chao Zhang, Behrend Heeren, Martin Rumpf, William A. P. Smith
In this paper we describe how to perform Principal Components Analysis in "shell space".
no code implementations • 8 Feb 2019 • Guangzhi Sun, Chao Zhang, Phil Woodland
This combination uses a 2-dimensional (2D) self-attentive structure, which extends the standard self-attentive layer by averaging not only across time but also across different types of embeddings.
no code implementations • 26 Mar 2019 • Chunzhi Gu, Xuequan Lu, Ying He, Chao Zhang
Complex blur such as the mixup of space-variant and space-invariant blur, which is hard to model mathematically, widely exists in real images.
no code implementations • 26 Mar 2019 • Takumi Nakane, Takuya Akashi, Xuequan Lu, Chao Zhang
We propose a novel genetic algorithm to solve the image deformation estimation problem by preserving the genetic diversity.
no code implementations • 2 Apr 2019 • Hang Zou, Chao Zhang, Samson Lasaulce
The proposed point of view is fully relevant for a receiver which has to send a quantized version of the channel state to the transmitter.
no code implementations • 11 Apr 2019 • Zhuo Lei, Chao Zhang, Qian Zhang, Guoping Qiu
In constructing the dataset, because of the subjectivity of user-generated video summarization, we manually annotate 25 summaries for each video, which are in total 1300 summaries.
no code implementations • 17 May 2019 • Hang Zou, Chao Zhang, Samson Lasaulce, Lucas Saludjian, Patrick Panciatici
In this paper, we introduce the problem of decision-oriented communications, that is, the goal of the source is to send the right amount of information in order for the intended destination to execute a task.
no code implementations • 8 Jun 2019 • Yu-cheng Chen, Matus Telgarsky, Chao Zhang, Bolton Bailey, Daniel Hsu, Jian Peng
This paper provides a simple procedure to fit generative networks to target distributions, with the goal of a small Wasserstein distance (or other optimal transport costs).
no code implementations • 18 Jun 2019 • Zhisheng Zhong, Fangyin Wei, Zhouchen Lin, Chao Zhang
Furthermore, we propose that weight tensors in networks with proper order and balanced dimension are easier to be compressed.
no code implementations • 23 Jun 2019 • Jian-Ya Ding, Chao Zhang, Lei Shen, Shengyin Li, Bing Wang, Yinghui Xu, Le Song
In many applications, a similar MIP model is solved on a regular basis, maintaining remarkable similarities in model structures and solution appearances but differing in formulation coefficients.
no code implementations • 21 Jun 2019 • Patrick von Platen, Chao Zhang, Philip Woodland
This paper proposes a novel multi-span structure for acoustic modelling based on the raw waveform with multiple streams of CNN input layers, each processing a different span of the raw waveform signal.
no code implementations • 2 Jul 2019 • Yi Zhang, Chao Zhang, Takuya Akashi
We propose a novel multi-scale template matching method which is robust against both scaling and rotation in unconstrained environments.
no code implementations • 16 Sep 2019 • Hang Zou, Chao Zhang, Samson Lasaulce, Lucas Saludjian, Patrick Panciatici
We propose a framework to find a good (finite) decision set which induces a minimal performance loss w. r. t.
no code implementations • 14 Sep 2019 • Qiujia Li, Chao Zhang, Philip C. Woodland
This paper proposes a novel automatic speech recognition (ASR) framework called Integrated Source-Channel and Attention (ISCA) that combines the advantages of traditional systems based on the noisy source-channel model (SC) and end-to-end style systems using attention-based sequence-to-sequence models.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
no code implementations • 8 Oct 2019 • Chao Zhang, Min-Hsiu Hsieh, DaCheng Tao
We also develop the tail inequalities for matrix random series and matrix martingale difference sequence.
no code implementations • 17 Oct 2019 • Jiaming Shen, Zeqiu Wu, Dongming Lei, Chao Zhang, Xiang Ren, Michelle T. Vanni, Brian M. Sadler, Jiawei Han
Taxonomies are of great value to many knowledge-rich applications.
no code implementations • 21 Oct 2019 • Jiahao Xie, Zebang Shen, Chao Zhang, Boyu Wang, Hui Qian
This paper focuses on projection-free methods for solving smooth Online Convex Optimization (OCO) problems.
no code implementations • 21 Oct 2019 • Chao Zhang, Jiahao Xie, Zebang Shen, Peilin Zhao, Tengfei Zhou, Hui Qian
In this paper, we explore a general Aggregated Gradient Langevin Dynamics framework (AGLD) for the Markov Chain Monte Carlo (MCMC) sampling.
no code implementations • 10 Nov 2019 • Chao Zhang, Zichao Yang, Xiaodong He, Li Deng
This review provides a comprehensive analysis of recent works on multimodal deep learning from three perspectives: learning multimodal representations, fusing multimodal signals at various levels, and multimodal applications.
no code implementations • 10 Nov 2019 • Yassir Fathullah, Chao Zhang, Philip C. Woodland
Speaker diarisation systems nowadays use embeddings generated from speech segments in a bottleneck layer, which are needed to be discriminative for unseen speakers.
no code implementations • 30 Sep 2019 • Yuan-Yuan Zhao, Chao Zhang, Shuming Cheng, Xinhui Li, Yu Guo, Bi-Heng Liu, Huan-Yu Ku, Shin-Liang Chen, Qiaoyan Wen, Yun-Feng Huang, Guo-Yong Xiang, Chuan-Feng Li, Guang-Can Guo
We first establish the DI verification framework, relying on the measurement-device-independent technique and self-testing, and show it is able to verify all EPR-steerable states.
Quantum Physics
no code implementations • 20 Jan 2020 • Chao Zhang, Xuequan Lu, Katsuya Hotta, Xi Yang
The WA data can be naturally obtained in an interactive way for specific tasks, for example, in the case of homography estimation, one can easily annotate points on the same plane/object with a single label by observing the image.
no code implementations • 30 Mar 2020 • Takumi Nakane, Xuequan Lu, Chao Zhang
In evolutionary algorithms, genetic operators iteratively generate new offspring which constitute a potentially valuable set of search history.
no code implementations • 13 Apr 2020 • Huajie Shao, Dachun Sun, Jiahao Wu, Zecheng Zhang, Aston Zhang, Shuochao Yao, Shengzhong Liu, Tianshi Wang, Chao Zhang, Tarek Abdelzaher
Motivated by this trend, we describe a novel item-item cross-platform recommender system, $\textit{paper2repo}$, that recommends relevant repositories on GitHub that match a given paper in an academic search system such as Microsoft Academic.
no code implementations • 1 May 2020 • Shi Zhi, Liyuan Liu, Yu Zhang, Shiyin Wang, Qi Li, Chao Zhang, Jiawei Han
While typical named entity recognition (NER) models require the training set to be annotated with all target types, each available datasets may only cover a part of them.
no code implementations • 11 Jul 2020 • Dongbo Zhang, Zheng Fang, Xuequan Lu, Hong Qin, Antonio Robles-Kelly, Chao Zhang, Ying He
3D human segmentation has seen noticeable progress in re-cent years.
no code implementations • 31 Aug 2020 • Chunzhi Gu, Xuequan Lu, Chao Zhang
In particular, we relate the transferred image with the example image under the Gaussian Mixture Model (GMM) and regard the transferred image color as the GMM centroids.
no code implementations • 12 Sep 2020 • Yi Zhou, Shuyang Sun, Chao Zhang, Yikang Li, Wanli Ouyang
By assigning each relationship a single label, current approaches formulate the relationship detection as a classification problem.
1 code implementation • 17 Mar 2019 • Xiangpan Ji, Wenqiang Gu, Xin Qian, Hanyu Wei, Chao Zhang
We describe an approximation to the widely-used Poisson-likelihood chi-square using a linear combination of Neyman's and Pearson's chi-squares, namely "combined Neyman-Pearson chi-square" ($\chi^2_{\mathrm{CNP}}$).
Data Analysis, Statistics and Probability High Energy Physics - Experiment Nuclear Experiment
no code implementations • 19 Oct 2020 • Hao Wang, Jia Zhang, Yingce Xia, Jiang Bian, Chao Zhang, Tie-Yan Liu
However, most existing studies overlook the code's intrinsic structural logic, which indeed contains a wealth of semantic information, and fails to capture intrinsic features of codes.
no code implementations • 21 Oct 2020 • Rui Feng, Jie Yuan, Chao Zhang
We argue that the event extraction models so trained are inherently label-hungry, and can generalize poorly across domains and text genres. We propose a reading comprehension framework for event extraction. Specifically, we formulate event detection as a textual entailment prediction problem, and argument detection as a question answer-ing problem.
no code implementations • 22 Oct 2020 • Guangzhi Sun, Chao Zhang, Phil Woodland
Significant progress has recently been made in speaker diarisation after the introduction of d-vectors as speaker embeddings extracted from neural network (NN) speaker classifiers for clustering speech segments.
no code implementations • 27 Oct 2020 • Wen Wu, Chao Zhang, Philip C. Woodland
In this paper, a novel two-branch neural network model structure is proposed for multimodal emotion recognition, which consists of a time synchronous branch (TSB) and a time asynchronous branch (TAB).
Automatic Speech Recognition Automatic Speech Recognition (ASR) +6
no code implementations • 22 Aug 2020 • Jinfeng Zeng, Chenfeng Cao, Chao Zhang, Pengxiang Xu, Bei Zeng
To obtain the full spectrum of the Hamiltonian, we use a quantum imaginary time evolution algorithm with high temperature, which prepares a thermal state with a small correlation length.
Quantum Physics
no code implementations • 6 Jun 2019 • Chuang-Shi Shen, Chao Zhang, Xiaosheng Gao, Yulong Li
We recognize that the fragmentation problem in shells is analogous to the cracking behavior of tree bark, and closed form solutions is obtained to describe the relationship between the expansion velocity and the number of necks with consideration of the strain rate dependent strength of the shell material.
Soft Condensed Matter Applied Physics
no code implementations • 6 Nov 2020 • Guanghui Xu, Wei Song, Zhengchen Zhang, Chao Zhang, Xiaodong He, BoWen Zhou
Despite prosody is related to the linguistic information up to the discourse structure, most text-to-speech (TTS) systems only take into account that within each sentence, which makes it challenging when converting a paragraph of texts into natural and expressive speech.
no code implementations • 13 Apr 2020 • Chao Zhang, Xiang Zhao, Kai Lin, Shaojun Zhang, Wen Zhao, Anzhong Wang
In particular, we find that, out of the five non-trivial field equations, only three are independent, so the problem is well-posed, as now generically there are only three unknown functions, {$F(r), B(r), A(r)$, where $F$ and $B$ are metric coefficients, and $A$ describes the aether field.}
General Relativity and Quantum Cosmology Astrophysics of Galaxies High Energy Physics - Phenomenology High Energy Physics - Theory
no code implementations • 2 Dec 2020 • Weijie Liu, Chao Zhang, Jiahao Xie, Zebang Shen, Hui Qian, Nenggan Zheng
Graph matching finds the correspondence of nodes across two graphs and is a basic task in graph-based machine learning.
no code implementations • 28 Jul 2020 • Wei Xue, Gang Quan, Chao Zhang, Guohong Ding, Xiaodong He, BoWen Zhou
Statistical signal processing based speech enhancement methods adopt expert knowledge to design the statistical models and linear filters, which is complementary to the deep neural network (DNN) based methods which are data-driven.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
no code implementations • 22 Dec 2020 • Samuel Yen-Chi Chen, Tzu-Chieh Wei, Chao Zhang, Haiwang Yu, Shinjae Yoo
This work presents a quantum convolutional neural network (QCNN) for the classification of high energy physics events.
no code implementations • 5 Jan 2021 • Chao Zhang, Joaquin Vanschoren, Arlette van Wissen, Daniel Lakens, Boris de Ruyter, Wijnand A. IJsselsteijn
Psychological theories of habit posit that when a strong habit is formed through behavioral repetition, it can trigger behavior automatically in the same environment.
no code implementations • 15 Jan 2021 • Samuel Yen-Chi Chen, Tzu-Chieh Wei, Chao Zhang, Haiwang Yu, Shinjae Yoo
This research provides a hybrid quantum-classical graph convolutional network (QGCNN) for learning HEP data.
no code implementations • 3 Mar 2021 • Chao Zhang, Wenqiang Yi, Yuanwei Liu, Qiang Wang
Numerical results indicate that 1) although the interference from other cells is enhanced via the RISs, the performance of the RIS-aided user still enhances since the channel quality is strengthened more obviously; and 2) the SIC order can be altered by employing the RISs since the RISs improve the channel quality of the aided user.
Information Theory Information Theory
no code implementations • 10 Mar 2021 • Chao Zhang, Shihan Wang, Henk Aarts, Mehdi Dastani
Reinforcement learning (RL) agents in human-computer interactions applications require repeated user interactions before they can perform well.
no code implementations • 12 Mar 2021 • Adnan Haider, Chao Zhang, Florian L. Kreyssig, Philip C. Woodland
This paper presents a novel natural gradient and Hessian-free (NGHF) optimisation framework for neural network training that can operate efficiently in a distributed manner.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
no code implementations • 1 Aug 2017 • Carl Yang, Lanxiao Bai, Chao Zhang, Quan Yuan, J. Han profile
In this work, we propose to devise a general and principled SSL (semi-supervised learning) framework, to alleviate data scarcity via smoothing among neighboring users and POIs, and treat various context by regularizing user preference based on context graphs.
no code implementations • 9 Apr 2021 • Chao Zhang, Xiaojun Chen, Shiqian Ma
In this paper, we propose a Riemannian smoothing steepest descent method to minimize a nonconvex and non-Lipschitz function on submanifolds.
no code implementations • 12 Apr 2021 • Xianjie Gao, Xueguan Song, Maolin Shi, Chao Zhang, Hongwei Zhang
In this paper, based on in-situ TBM operational data, we use the machine-learning (ML) methods to build the real-time forecast models for TBM load parameters, which can instantaneously provide the future values of the TBM load parameters as long as the current data are collected.
no code implementations • 12 Apr 2021 • Samson Lasaulce, Chao Zhang, Vineeth Varma, Irinel Constantin Morarescu
Should the measures be more (or less) restrictive?
no code implementations • 23 Apr 2021 • Yi He, Haoran Xie, Chao Zhang, Xi Yang, Kazunori Miyata
This paper proposes a deep generative model for generating normal maps from users sketch with geometric sampling.
no code implementations • 23 Apr 2021 • Sheldon Fung, Xuequan Lu, Chao Zhang, Chang-Tsun Li
Extensive experiments show that our unsupervised learning method enables comparable detection performance to state-of-the-art supervised techniques, in both the intra- and inter-dataset settings.
1 code implementation • 8 May 2021 • Lingwei Peng, Hui Qian, Zhebang Shen, Chao Zhang, Fei Li
Model-free deep reinforcement learning has achieved great success in many domains, such as video games, recommendation systems and robotic control tasks.
no code implementations • 2 Jun 2021 • Chao Zhang, Samson Lasaulce, Martin Hennebel, Lucas Saludjian, Patrick Panciatici, H. Vincent Poor
For this purpose, we formulate the framework of decision-making oriented clustering and propose an algorithm providing a decision-based partition of the data space and good representative decisions.