9 code implementations • NeurIPS 2019 • Li Dong, Nan Yang, Wenhui Wang, Furu Wei, Xiaodong Liu, Yu Wang, Jianfeng Gao, Ming Zhou, Hsiao-Wuen Hon
This paper presents a new Unified pre-trained Language Model (UniLM) that can be fine-tuned for both natural language understanding and generation tasks.
Ranked #2 on Generative Question Answering on CoQA (using extra training data)
3 code implementations • 28 Feb 2020 • Hangbo Bao, Li Dong, Furu Wei, Wenhui Wang, Nan Yang, Xiaodong Liu, Yu Wang, Songhao Piao, Jianfeng Gao, Ming Zhou, Hsiao-Wuen Hon
We propose to pre-train a unified language model for both autoencoding and partially autoregressive language modeling tasks using a novel training procedure, referred to as a pseudo-masked language model (PMLM).
Ranked #4 on Question Generation on SQuAD1.1 (using extra training data)
1 code implementation • ICML 2020 • Hangbo Bao, Li Dong, Furu Wei, Wenhui Wang, Nan Yang, Xiaodong Liu, Yu Wang, Jianfeng Gao, Songhao Piao, Ming Zhou, Hsiao-Wuen Hon
We propose to pre-train a unified language model for both autoencoding and partially autoregressive language modeling tasks using a novel training procedure, referred to as a pseudo-masked language model (PMLM).
2 code implementations • 15 Nov 2022 • Yu Wang, Xin Li, Shengzhao Wen, Fukui Yang, Wanping Zhang, Gang Zhang, Haocheng Feng, Junyu Han, Errui Ding
In this paper, we focus on the compression of DETR with knowledge distillation.
1 code implementation • 19 Jan 2020 • Yi Wang, Yang Yang, Weiguo Zhu, Yi Wu, Xu Yan, Yongfeng Liu, Yu Wang, Liang Xie, Ziyao Gao, Wenjing Zhu, Xiang Chen, Wei Yan, Mingjie Tang, Yuan Tang
Previous database systems extended their SQL dialect to support ML.
7 code implementations • 7 May 2019 • Yu Wang, Quan Zhou, Jia Liu, Jian Xiong, Guangwei Gao, Xiaofu Wu, Longin Jan Latecki
LEDNet: A Lightweight Encoder-Decoder Network for Real-time Semantic Segmentation
Ranked #29 on Real-Time Semantic Segmentation on Cityscapes test
2 code implementations • 24 Jun 2019 • Yu Wang, Quan Zhou, Xiaofu Wu
The whole network has nearly symmetric architecture, which is mainly composed of a series of factorized convolution unit (FCU) and its parallel counterparts (PFCU).
Ranked #28 on Real-Time Semantic Segmentation on Cityscapes test
15 code implementations • 2 Mar 2021 • Chao Yu, Akash Velu, Eugene Vinitsky, Jiaxuan Gao, Yu Wang, Alexandre Bayen, Yi Wu
This is often due to the belief that PPO is significantly less sample efficient than off-policy methods in multi-agent systems.
Multi-agent Reinforcement Learning reinforcement-learning +3
3 code implementations • ACL 2020 • Xiaodong Liu, Yu Wang, Jianshu ji, Hao Cheng, Xueyun Zhu, Emmanuel Awa, Pengcheng He, Weizhu Chen, Hoifung Poon, Guihong Cao, Jianfeng Gao
We present MT-DNN, an open-source natural language understanding (NLU) toolkit that makes it easy for researchers and developers to train customized deep learning models.
3 code implementations • 20 Apr 2020 • Xiaodong Liu, Hao Cheng, Pengcheng He, Weizhu Chen, Yu Wang, Hoifung Poon, Jianfeng Gao
In natural language processing (NLP), pre-training large neural language models such as BERT have demonstrated impressive gain in generalization for a variety of tasks, with further improvement from adversarial fine-tuning.
Ranked #6 on Natural Language Inference on ANLI test (using extra training data)
2 code implementations • 7 Jul 2020 • Guyue Huang, Guohao Dai, Yu Wang, Huazhong Yang
GE-SpMM performs SpMM-like operation on sparse matrices represented in the most common Compressed Sparse Row (CSR) format, so it can be embedded in GNN frameworks with no preprocessing overheads and support general GNN algorithms.
Distributed, Parallel, and Cluster Computing
1 code implementation • 1 Mar 2021 • Yukuo Cen, Zhenyu Hou, Yan Wang, Qibin Chen, Yizhen Luo, Zhongming Yu, Hengrui Zhang, Xingcheng Yao, Aohan Zeng, Shiguang Guo, Yuxiao Dong, Yang Yang, Peng Zhang, Guohao Dai, Yu Wang, Chang Zhou, Hongxia Yang, Jie Tang
In CogDL, we propose a unified design for the training and evaluation of GNN models for various graph tasks, making it unique among existing graph learning libraries.
1 code implementation • 20 Feb 2019 • Yu Xing, Shuang Liang, Lingzhi Sui, Xijie Jia, Jiantao Qiu, Xin Liu, Yushun Wang, Yu Wang, Yi Shan
On the Xilinx ZU2 @330 MHz and ZU9 @330 MHz, we achieve equivalently state-of-the-art performance on our benchmarks by na\"ive implementations without optimizations, and the throughput is further improved up to 1. 26x by leveraging heterogeneous optimizations in DNNVM.
1 code implementation • 25 Oct 2023 • Haotian Tang, Shang Yang, Zhijian Liu, Ke Hong, Zhongming Yu, Xiuyu Li, Guohao Dai, Yu Wang, Song Han
On top of this, we design the Sparse Autotuner, which extends the design space of existing sparse convolution libraries and searches for the best dataflow configurations for training and inference workloads.
1 code implementation • 13 Jul 2020 • Ahmed Elnaggar, Michael Heinzinger, Christian Dallago, Ghalia Rihawi, Yu Wang, Llion Jones, Tom Gibbs, Tamas Feher, Christoph Angerer, Martin Steinegger, Debsindhu Bhowmik, Burkhard Rost
Here, we trained two auto-regressive models (Transformer-XL, XLNet) and four auto-encoder models (BERT, Albert, Electra, T5) on data from UniRef and BFD containing up to 393 billion amino acids.
Ranked #1 on Protein Secondary Structure Prediction on CASP12
Dimensionality Reduction Protein Secondary Structure Prediction
1 code implementation • 10 Feb 2022 • Benedek Rozemberczki, Charles Tapley Hoyt, Anna Gogleva, Piotr Grabowski, Klas Karis, Andrej Lamov, Andriy Nikolov, Sebastian Nilsson, Michael Ughetto, Yu Wang, Tyler Derr, Benjamin M Gyori
In this paper, we introduce ChemicalX, a PyTorch-based deep learning library designed for providing a range of state of the art models to solve the drug pair scoring task.
2 code implementations • 13 Mar 2024 • Yusheng Liao, Yutong Meng, Yuhao Wang, Hongcheng Liu, Yanfeng Wang, Yu Wang
Large Language Models (LLMs) have demonstrated remarkable proficiency in human interactions, yet their application within the medical field remains insufficiently explored.
2 code implementations • 13 Apr 2024 • Yusheng Liao, Shuyang Jiang, Yu Wang, Yanfeng Wang
Large language models like ChatGPT have shown substantial progress in natural language understanding and generation, proving valuable across various disciplines, including the medical field.
1 code implementation • CVPR 2021 • Jialian Wu, Jiale Cao, Liangchen Song, Yu Wang, Ming Yang, Junsong Yuan
Most online multi-object trackers perform object detection stand-alone in a neural net without any input from tracking.
Ranked #1 on Instance Segmentation on nuScenes
1 code implementation • ECCV 2020 • Tong Wu, Qingqiu Huang, Ziwei Liu, Yu Wang, Dahua Lin
We present a new loss function called Distribution-Balanced Loss for the multi-label recognition problems that exhibit long-tailed class distributions.
Ranked #7 on Long-tail Learning on VOC-MLT
1 code implementation • 12 Sep 2022 • Zixiang Zhou, Xiangchen Zhao, Yu Wang, Panqu Wang, Hassan Foroosh
It then uses the feature of the center candidate as the query embedding in the transformer.
Ranked #2 on 3D Object Detection on waymo cyclist
3 code implementations • ICLR 2018 • Yujun Lin, Song Han, Huizi Mao, Yu Wang, William J. Dally
The situation gets even worse with distributed training on mobile devices (federated learning), which suffers from higher latency, lower throughput, and intermittent poor connections.
1 code implementation • ECCV 2020 • Xuefei Ning, Yin Zheng, Tianchen Zhao, Yu Wang, Huazhong Yang
Experimental results on various search spaces confirm GATES's effectiveness in improving the performance predictor.
1 code implementation • NeurIPS 2021 • Xuefei Ning, Changcheng Tang, Wenshuo Li, Zixuan Zhou, Shuang Liang, Huazhong Yang, Yu Wang
Conducting efficient performance estimations of neural architectures is a major challenge in neural architecture search (NAS).
1 code implementation • 25 Nov 2020 • Xuefei Ning, Changcheng Tang, Wenshuo Li, Songyi Yang, Tianchen Zhao, Niansong Zhang, Tianyi Lu, Shuang Liang, Huazhong Yang, Yu Wang
Neural Architecture Search (NAS) has received extensive attention due to its capability to discover neural network architectures in an automated manner.
1 code implementation • 22 Dec 2020 • Xuefei Ning, Junbo Zhao, Wenshuo Li, Tianchen Zhao, Yin Zheng, Huazhong Yang, Yu Wang
In this paper, considering scenarios with capacity budget, we aim to discover adversarially robust architecture at targeted capacities.
1 code implementation • 16 Jul 2022 • Zixuan Zhou, Xuefei Ning, Yi Cai, Jiashu Han, Yiping Deng, Yuhan Dong, Huazhong Yang, Yu Wang
Specifically, we train the supernet with a large sharing extent (an easier curriculum) at the beginning and gradually decrease the sharing extent of the supernet (a harder curriculum).
1 code implementation • 16 May 2020 • Nick Altieri, Rebecca L. Barter, James Duncan, Raaz Dwivedi, Karl Kumbier, Xiao Li, Robert Netzorg, Briton Park, Chandan Singh, Yan Shuo Tan, Tiffany Tang, Yu Wang, Chao Zhang, Bin Yu
We use this data to develop predictions and corresponding prediction intervals for the short-term trajectory of COVID-19 cumulative death counts at the county-level in the United States up to two weeks ahead.
1 code implementation • The International Conference on Learning Representations 2017 • Yujun Lin, Song Han, Huizi Mao, Yu Wang, W. Dally
Large-scale distributed training requires significant communication bandwidth for gradient exchange that limits the scalability of multi-node training, and requires expensive high-bandwidth network infrastructure.
2 code implementations • 29 Jul 2018 • Xiujun Li, Yu Wang, Siqi Sun, Sarah Panda, Jingjing Liu, Jianfeng Gao
This proposal introduces a Dialogue Challenge for building end-to-end task-completion dialogue systems, with the goal of encouraging the dialogue research community to collaborate and benchmark on standard datasets and unified experimental environment.
1 code implementation • 19 Dec 2023 • Weipeng Guan, Peiyu Chen, Huibin Zhao, Yu Wang, Peng Lu
To the best of our knowledge, this is the first non-learning work to realize event-based dense mapping.
1 code implementation • 22 Aug 2023 • Yu Wang, Nedim Lipka, Ryan A. Rossi, Alexa Siu, Ruiyi Zhang, Tyler Derr
Concurrently, the graph traversal agent acts as a local navigator that gathers pertinent context to progressively approach the question and guarantee retrieval quality.
6 code implementations • 23 Jun 2020 • Runzhou Ge, Zhuangzhuang Ding, Yihan Hu, Yu Wang, Sijia Chen, Li Huang, Yuan Li
High-efficiency point cloud 3D object detection operated on embedded systems is important for many robotics applications including autonomous driving.
1 code implementation • 11 Jul 2022 • Ting Yao, Yehao Li, Yingwei Pan, Yu Wang, Xiao-Ping Zhang, Tao Mei
Dual-ViT is henceforth able to reduce the computational complexity without compromising much accuracy.
2 code implementations • 2 Jan 2022 • Huaxiu Yao, Yu Wang, Sai Li, Linjun Zhang, Weixin Liang, James Zou, Chelsea Finn
Machine learning algorithms typically assume that training and test examples are drawn from the same distribution.
1 code implementation • 28 Jul 2023 • Xuefei Ning, Zinan Lin, Zixuan Zhou, Zifu Wang, Huazhong Yang, Yu Wang
This work aims at decreasing the end-to-end generation latency of large language models (LLMs).
1 code implementation • 22 Sep 2023 • Botian Xu, Feng Gao, Chao Yu, Ruize Zhang, Yi Wu, Yu Wang
In this work, we introduce OmniDrones, an efficient and flexible platform tailored for reinforcement learning in drone control, built on Nvidia's Omniverse Isaac Sim.
1 code implementation • CVPR 2021 • Tong Wu, Ziwei Liu, Qingqiu Huang, Yu Wang, Dahua Lin
We then perform a systematic study on existing long-tailed recognition methods in conjunction with the adversarial training framework.
3 code implementations • 2 Mar 2023 • Sheng Zhang, Yanbo Xu, Naoto Usuyama, Hanwen Xu, Jaspreet Bagga, Robert Tinn, Sam Preston, Rajesh Rao, Mu Wei, Naveen Valluri, Cliff Wong, Andrea Tupini, Yu Wang, Matt Mazzola, Swadheen Shukla, Lars Liden, Jianfeng Gao, Matthew P. Lungren, Tristan Naumann, Sheng Wang, Hoifung Poon
Therefore, training an effective generalist biomedical model requires high-quality multimodal data, such as parallel image-text pairs.
Ranked #3 on Medical Visual Question Answering on SLAKE-English
1 code implementation • 30 Oct 2023 • Qiao Sun, Shiduo Zhang, Danjiao Ma, Jingzhe Shi, Derun Li, Simian Luo, Yu Wang, Ningyi Xu, Guangzhi Cao, Hang Zhao
STR reformulates the motion prediction and motion planning problems by arranging observations, states, and actions into one unified sequence modeling task.
1 code implementation • CVPR 2020 • Qi Cai, Yingwei Pan, Yu Wang, Jingen Liu, Ting Yao, Tao Mei
To this end, we devise a general loss function to cover most region-based object detectors with various sampling strategies, and then based on it we propose a unified sample weighting network to predict a sample's task weights.
1 code implementation • 22 Sep 2021 • Fan Zhang, Yu Wang, Hua Yang
For the context block, we propose strip pooling module to capture anisotropic and long-range contextual information, which exists in abdominal scene.
1 code implementation • 15 Jun 2023 • Ziyang Ma, Zhisheng Zheng, Guanrou Yang, Yu Wang, Chao Zhang, Xie Chen
Our models outperform other SSL models significantly on the LibriSpeech benchmark without the need for iterative re-clustering and re-training.
1 code implementation • 19 Mar 2022 • Junwen Pan, Pengfei Zhu, Kaihua Zhang, Bing Cao, Yu Wang, Dingwen Zhang, Junwei Han, QinGhua Hu
Semantic segmentation with limited annotations, such as weakly supervised semantic segmentation (WSSS) and semi-supervised semantic segmentation (SSSS), is a challenging task that has attracted much attention recently.
Ranked #34 on Weakly-Supervised Semantic Segmentation on COCO 2014 val
1 code implementation • 28 Feb 2024 • Shiyao Li, Xuefei Ning, Luning Wang, Tengxuan Liu, Xiangsheng Shi, Shengen Yan, Guohao Dai, Huazhong Yang, Yu Wang
Post-training quantization (PTQ) has emerged as a promising technique to reduce the cost of large language models (LLMs).
2 code implementations • 6 Aug 2019 • Pengfei Zhu, Xinjie Yao, Yu Wang, Binyuan Hui, Dawei Du, QinGhua Hu
Dnet learns view-specific self-representation matrices, whereas Unet learns a common self-representation matrix for all views.
Ranked #1 on Multi-view Subspace Clustering on ORL
1 code implementation • 3 Apr 2021 • Yu Wang, Chee Siang Leow, Akio Kobayashi, Takehito Utsuro, Hiromitsu Nishizaki
This paper describes the ExKaldi-RT online automatic speech recognition (ASR) toolkit that is implemented based on the Kaldi ASR toolkit and Python language.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
1 code implementation • 6 Oct 2022 • Yu Wang, Chao Pang, Yuzhe Wang, Yi Jiang, Junru Jin, Sirui Liang, Quan Zou, Leyi Wei
Leveraging artificial intelligence for automatic retrosynthesis speeds up organic pathway planning in digital laboratories.
1 code implementation • 2 May 2021 • Zhiwei Liu, Ziwei Fan, Yu Wang, Philip S. Yu
We firstly pre-train a transformer with sequences in a reverse direction to predict prior items.
1 code implementation • 3 Jul 2022 • Yu Wang, Yuying Zhao, Yi Zhang, Tyler Derr
Graph Neural Networks (GNNs) have been successfully adopted in recommender systems by virtue of the message-passing that implicitly captures collaborative effect.
2 code implementations • 13 Mar 2021 • Shaowei Chen, Yu Wang, Jie Liu, Yuelin Wang
Aspect sentiment triplet extraction (ASTE), which aims to identify aspects from review sentences along with their corresponding opinion expressions and sentiments, is an emerging task in fine-grained opinion mining.
Aspect Sentiment Triplet Extraction Machine Reading Comprehension +2
2 code implementations • 1 Dec 2021 • Yu Wang, Yuying Zhao, Neil Shah, Tyler Derr
To this end, we introduce a novel framework, Graph-of-Graph Neural Networks (G$^2$GNN), which alleviates the graph imbalance issue by deriving extra supervision globally from neighboring graphs and locally from stochastic augmentations of graphs.
1 code implementation • NeurIPS 2020 • Qi Cai, Yu Wang, Yingwei Pan, Ting Yao, Tao Mei
This paper explores useful modifications of the recent development in contrastive learning via novel probabilistic modeling.
2 code implementations • ICLR 2021 • Zhenggang Tang, Chao Yu, Boyuan Chen, Huazhe Xu, Xiaolong Wang, Fei Fang, Simon Du, Yu Wang, Yi Wu
We propose a simple, general and effective technique, Reward Randomization for discovering diverse strategic policies in complex multi-agent games.
1 code implementation • 9 Oct 2023 • Haoyu Zhang, Yu Wang, Guanghao Yin, Kejun Liu, Yuanyuan Liu, Tianshu Yu
Though Multimodal Sentiment Analysis (MSA) proves effective by utilizing rich information from multiple sources (e. g., language, video, and audio), the potential sentiment-irrelevant and conflicting information across modalities may hinder the performance from being further improved.
Ranked #1 on Multimodal Sentiment Analysis on CMU-MOSEI (Acc-7 metric)
1 code implementation • 10 Jan 2021 • Guyue Huang, Jingbo Hu, Yifan He, Jialong Liu, Mingyuan Ma, Zhaoyang Shen, Juejian Wu, Yuanfan Xu, Hengrui Zhang, Kai Zhong, Xuefei Ning, Yuzhe ma, HaoYu Yang, Bei Yu, Huazhong Yang, Yu Wang
With the down-scaling of CMOS technology, the design complexity of very large-scale integrated (VLSI) is increasing.
1 code implementation • NAACL 2018 • Yu Wang, Yilin Shen, Hongxia Jin
The most effective algorithms are based on the structures of sequence to sequence models (or "encoder-decoder" models), and generate the intents and semantic tags either using separate models or a joint model.
Ranked #1 on Intent Detection on ATIS
1 code implementation • Pattern Recognition 2023 • Dichao Liu, Longjiao Zhao, Yu Wang, Jien Kato
Specifically, this work views the shallow to deep layers of CNNs as “experts” knowledgeable about different perspectives.
Ranked #1 on Fine-Grained Image Classification on Stanford Cars (using extra training data)
1 code implementation • EMNLP 2018 • Bailin Wang, Wei Lu, Yu Wang, Hongxia Jin
It is common that entity mentions can contain other mentions recursively.
Ranked #6 on Nested Named Entity Recognition on NNE
Nested Mention Recognition Nested Named Entity Recognition +1
2 code implementations • 14 Jul 2022 • Yu Wang, Guan Gui, Yun Lin, Hsiao-Chun Wu, Chau Yuen, Fumiyuki Adachi
Thus, we focus on few-shot SEI (FS-SEI) for aircraft identification via automatic dependent surveillance-broadcast (ADS-B) signals, and a novel FS-SEI method is proposed, based on deep metric ensemble learning (DMEL).
1 code implementation • 22 Oct 2021 • Yu Wang, Charu Aggarwal, Tyler Derr
Recent years have witnessed the significant success of applying graph neural networks (GNNs) in learning effective node representations for classification.
1 code implementation • 27 Aug 2022 • Yu Wang, Hengrui Zhang, Zhiwei Liu, Liangwei Yang, Philip S. Yu
Then we propose Contrastive Variational AutoEncoder (ContrastVAE in short), a two-branched VAE model with contrastive regularization as an embodiment of ContrastELBO for sequential recommendation.
1 code implementation • ICCV 2023 • An Yan, Yu Wang, Yiwu Zhong, chengyu dong, Zexue He, Yujie Lu, William Wang, Jingbo Shang, Julian McAuley
Recent advances in foundation models present new opportunities for interpretable visual recognition -- one can first query Large Language Models (LLMs) to obtain a set of attributes that describe each class, then apply vision-language models to classify images via these attributes.
1 code implementation • 20 Aug 2023 • Zihan Zhao, Yiyang Jiang, Heyang Liu, Yanfeng Wang, Yu Wang
While Large Language Models (LLMs) have demonstrated commendable performance across a myriad of domains and tasks, existing LLMs still exhibit a palpable deficit in handling multimodal functionalities, especially for the Spoken Question Answering (SQA) task which necessitates precise alignment and deep interaction between speech and text features.
1 code implementation • ACL 2020 • Shaowei Chen, Jie Liu, Yu Wang, Wenzheng Zhang, Ziming Chi
The opinion entity extraction unit and the relation detection unit are developed as two channels to extract opinion entities and relations simultaneously.
1 code implementation • NeurIPS 2020 • Hong Liu, Mingsheng Long, Jianmin Wang, Yu Wang
(2) Since the target data arrive online, the agent should also maintain competence on previous target domains, i. e. to adapt without forgetting.
1 code implementation • 3 Feb 2023 • Chao Yu, Jiaxuan Gao, Weilin Liu, Botian Xu, Hao Tang, Jiaqi Yang, Yu Wang, Yi Wu
A crucial limitation of this framework is that every policy in the pool is optimized w. r. t.
1 code implementation • ICCV 2023 • Yichen Yuan, Yifan Wang, Lijun Wang, Xiaoqi Zhao, Huchuan Lu, Yu Wang, Weibo Su, Lei Zhang
Recent leading zero-shot video object segmentation (ZVOS) works devote to integrating appearance and motion information by elaborately designing feature fusion modules and identically applying them in multiple feature stages.
1 code implementation • 7 Jun 2022 • Yu Wang, Yuying Zhao, Yushun Dong, Huiyuan Chen, Jundong Li, Tyler Derr
Motivated by our analysis, we propose Fair View Graph Neural Network (FairVGNN) to generate fair views of features by automatically identifying and masking sensitive-correlated features considering correlation variation after feature propagation.
1 code implementation • 2 Feb 2024 • Kun-Peng Ning, Shuo Yang, Yu-Yang Liu, Jia-Yu Yao, Zhen-Hui Liu, Yu Wang, Ming Pang, Li Yuan
Existing large language models (LLMs) evaluation methods typically focus on testing the performance on some closed-environment and domain-specific benchmarks with human annotations.
1 code implementation • 6 Feb 2024 • Tao Yuan, Xuefei Ning, Dong Zhou, Zhijie Yang, Shiyao Li, Minghui Zhuang, Zheyue Tan, Zhuyu Yao, Dahua Lin, Boxun Li, Guohao Dai, Shengen Yan, Yu Wang
In contrast, the average context lengths of mainstream benchmarks are insufficient (5k-21k), and they suffer from potential knowledge leakage and inaccurate metrics, resulting in biased evaluation.
2 code implementations • 16 Sep 2019 • Alper Kamil Bozkurt, Yu Wang, Michael M. Zavlanos, Miroslav Pajic
We present a reinforcement learning (RL) framework to synthesize a control policy from a given linear temporal logic (LTL) specification in an unknown stochastic environment that can be modeled as a Markov Decision Process (MDP).
1 code implementation • 17 Mar 2022 • Kai Zhang, Yu Wang, Hongyi Wang, Lifu Huang, Carl Yang, Xun Chen, Lichao Sun
Furthermore, we propose a Federated learning paradigm with privacy-preserving Relation embedding aggregation (FedR) to tackle the privacy issue in FedE.
1 code implementation • 14 Mar 2024 • Yu Wang, Xiaogeng Liu, Yu Li, Muhao Chen, Chaowei Xiao
However, with the integration of additional modalities, MLLMs are exposed to new vulnerabilities, rendering them prone to structured-based jailbreak attacks, where semantic content (e. g., "harmful text") has been injected into the images to mislead MLLMs.
1 code implementation • 27 Mar 2024 • Qiran Zou, Shangyuan Yuan, Shian Du, Yu Wang, Chang Liu, Yi Xu, Jie Chen, Xiangyang Ji
However, these methods encounter challenges such as the lack of coordination between different part motions and difficulties for networks to understand part concepts.
Ranked #8 on Motion Synthesis on HumanML3D
1 code implementation • 22 Oct 2020 • Tao Lu, Yuanzhi Wang, Yanduo Zhang, Yu Wang, Wei Liu, Zhongyuan Wang, Junjun Jiang
However, most of them fail to take into account the overall facial profile and fine texture details simultaneously, resulting in reduced naturalness and fidelity of the reconstructed face, and further impairing the performance of downstream tasks (e. g., face detection, facial recognition).
1 code implementation • 26 Aug 2021 • Yu Wang, Zhiwei Liu, Ziwei Fan, Lichao Sun, Philip S. Yu
In the information explosion era, recommender systems (RSs) are widely studied and applied to discover user-preferred information.
1 code implementation • CVPR 2023 • Yu Wang, Pengchong Qiao, Chang Liu, Guoli Song, Xiawu Zheng, Jie Chen
We argue that an overlooked problem of robust SSL is its corrupted information on semantic level, practically limiting the development of the field.
1 code implementation • 25 Aug 2021 • Yu Wang, Tyler Derr
Nevertheless, iterative propagation restricts the information of higher-layer neighborhoods to be transported through and fused with the lower-layer neighborhoods', which unavoidably results in feature smoothing between neighborhoods in different layers and can thus compromise the performance, especially on heterophily networks.
Ranked #25 on Node Classification on Cora
1 code implementation • 15 Jun 2023 • Enshu Liu, Xuefei Ning, Zinan Lin, Huazhong Yang, Yu Wang
Diffusion probabilistic models (DPMs) are a new class of generative models that have achieved state-of-the-art generation quality in various domains.
1 code implementation • NeurIPS 2021 • Jiayu Chen, Yuanxin Zhang, Yuanfan Xu, Huimin Ma, Huazhong Yang, Jiaming Song, Yu Wang, Yi Wu
We motivate our paradigm through a variational perspective, where the learning objective can be decomposed into two terms: task learning on the current task distribution, and curriculum update to a new task distribution.
1 code implementation • 31 Aug 2023 • Yi Zhang, Yuying Zhao, Zhaoqing Li, Xueqi Cheng, Yu Wang, Olivera Kotevska, Philip S. Yu, Tyler Derr
Despite this progress, there is a lack of a comprehensive overview of the attacks and the techniques for preserving privacy in the graph domain.
2 code implementations • NeurIPS 2021 • Huaxiu Yao, Yu Wang, Ying WEI, Peilin Zhao, Mehrdad Mahdavi, Defu Lian, Chelsea Finn
In ATS, for the first time, we design a neural scheduler to decide which meta-training tasks to use next by predicting the probability being sampled for each candidate task, and train the scheduler to optimize the generalization capacity of the meta-model to unseen tasks.
1 code implementation • 8 May 2021 • JiaMing Wang, Zhenfeng Shao, Tao Lu, Xiao Huang, Ruiqian Zhang, Yu Wang
Despite their success, however, low/high spatial resolution pairs are usually difficult to obtain in satellites with a high temporal resolution, making such approaches in SR impractical to use.
4 code implementations • 14 May 2019 • Weitian Li, Haiguang Xu, Zhixian Ma, Dan Hu, Zhenghao Zhu, Chenxi Shan, Jingying Wang, Junhua Gu, Dongchao Zheng, Xiaoli Lian, Qian Zheng, Yu Wang, Jie Zhu, Xiang-Ping Wu
The overwhelming foreground contamination is one of the primary impediments to probing the EoR through measuring the redshifted 21 cm signal.
Cosmology and Nongalactic Astrophysics
1 code implementation • 9 Aug 2022 • Lantu Guo, Yu Wang, Yun Lin, Haitao Zhao, Guan Gui
Automatic modulation classification (AMC) is a key technique for designing non-cooperative communication systems, and deep learning (DL) is applied effectively to AMC for improving classification accuracy.
1 code implementation • 16 Aug 2023 • Tianyu Fu, Chiyue Wei, Yu Wang, Rex Ying
We introduce DeSCo, a scalable neural deep subgraph counting pipeline, designed to accurately predict both the count and occurrence position of queries on target graphs post single training.
1 code implementation • 7 Feb 2022 • Liangwei Yang, Zhiwei Liu, Yu Wang, Chen Wang, Ziwei Fan, Philip S. Yu
We conduct a comprehensive analysis of users' online game behaviors, which motivates the necessity of handling those three characteristics in the online game recommendation.
1 code implementation • 10 Jul 2023 • Yu Wang, Emma R. Cobian, Jubilee Lee, Fang Liu, Jonathan D. Hauenstein, Daniele E. Schiavazzi
Variational inference is an increasingly popular method in statistics and machine learning for approximating probability distributions.
1 code implementation • 22 Aug 2023 • Mohamed Elaraby, Mengyin Lu, Jacob Dunn, Xueying Zhang, Yu Wang, Shizhu Liu, Pingchuan Tian, Yuping Wang, Yuxuan Wang
Large Language Models (LLMs) have revolutionized Natural Language Processing (NLP).
3 code implementations • NeurIPS 2019 • Xiao Li, Yu Wang, Sumanta Basu, Karl Kumbier, Bin Yu
Based on the original definition of MDI by Breiman et al. for a single tree, we derive a tight non-asymptotic bound on the expected bias of MDI importance of noisy features, showing that deep trees have higher (expected) feature selection bias than shallow ones.
1 code implementation • 26 Apr 2022 • Yu Wang, Yu Dong, Xuesong Lu, Aoying Zhou
Current deep learning models for code summarization generally follow the principle in neural machine translation and adopt the encoder-decoder framework, where the encoder learns the semantic representations from source code and the decoder transforms the learnt representations into human-readable text that describes the functionality of code snippets.
1 code implementation • 23 Dec 2023 • Jijia Liu, Chao Yu, Jiaxuan Gao, Yuqing Xie, Qingmin Liao, Yi Wu, Yu Wang
AI agents powered by Large Language Models (LLMs) have made significant advances, enabling them to assist humans in diverse complex tasks and leading to a revolution in human-AI coordination.
1 code implementation • 21 Jul 2017 • Yu Wang, Mirela Ben-Chen, Iosif Polterovich, Justin Solomon
We propose using the Dirichlet-to-Neumann operator as an extrinsic alternative to the Laplacian for spectral geometry processing and shape analysis.
Graphics
1 code implementation • 15 Nov 2022 • Zhaofan Qiu, Yehao Li, Yu Wang, Yingwei Pan, Ting Yao, Tao Mei
In this paper, we propose a novel deep architecture tailored for 3D point cloud applications, named as SPE-Net.
1 code implementation • 12 Jan 2024 • Yu Wang, Junxian Mu, Pengfei Zhu, QinGhua Hu
We show that the differences in attention maps can lead to diverse representations so that the fused representations can well handle the open space.
1 code implementation • 8 Nov 2023 • Ruochi Zhang, Haoran Wu, Yuting Xiu, Kewei Li, Ningning Chen, Yu Wang, Yan Wang, Xin Gao, Fengfeng Zhou
In recent years, the scientific community has become increasingly interested on peptides with non-canonical amino acids due to their superior stability and resistance to proteolytic degradation.
1 code implementation • 12 Jul 2019 • Yu Wang, Fengjuan Gao, Linzhang Wang, Ke Wang
In a cross-project prediction task, three neural bug detectors we instantiate from NeurSA are effective in catching null pointer dereference, array index out of bound and class cast bugs in unseen code.
1 code implementation • 19 Feb 2024 • Yu Wang, Zeyuan Zhang, Julian McAuley, Zexue He
To address this issue, we propose Long Video Chat (LVChat), where Frame-Scalable Encoding (FSE) is introduced to dynamically adjust the number of embeddings in alignment with the duration of the video to ensure long videos are not overly compressed into a few embeddings.
1 code implementation • ECCV 2020 • Xuefei Ning, Tianchen Zhao, Wenshuo Li, Peng Lei, Yu Wang, Huazhong Yang
In budgeted pruning, how to distribute the resources across layers (i. e., sparsity allocation) is the key problem.
1 code implementation • 13 Jun 2021 • Nianhui Guo, Joseph Bethge, Haojin Yang, Kai Zhong, Xuefei Ning, Christoph Meinel, Yu Wang
Recent works on Binary Neural Networks (BNNs) have made promising progress in narrowing the accuracy gap of BNNs to their 32-bit counterparts.
1 code implementation • 26 Sep 2022 • Jingyang Lin, Yu Wang, Qi Cai, Yingwei Pan, Ting Yao, Hongyang Chao, Tao Mei
Existing works attempt to solve the problem by explicitly imposing uncertainty on classifiers when OOD inputs are exposed to the classifier during training.
2 code implementations • 9 Jan 2023 • Chao Yu, Xinyi Yang, Jiaxuan Gao, Jiayu Chen, Yunfei Li, Jijia Liu, Yunfei Xiang, Ruixin Huang, Huazhong Yang, Yi Wu, Yu Wang
Simply waiting for every robot being ready for the next action can be particularly time-inefficient.
Multi-agent Reinforcement Learning reinforcement-learning +1
1 code implementation • 15 Jan 2024 • Zhilin Huang, Ling Yang, Zaixi Zhang, Xiangxin Zhou, Yu Bao, Xiawu Zheng, Yuwei Yang, Yu Wang, Wenming Yang
Then the selected protein-ligand subcomplex is processed with SE(3)-equivariant neural networks, and transmitted back to each atom of the complex for augmenting the target-aware 3D molecule diffusion generation with binding interaction information.
1 code implementation • 24 Jun 2022 • Yushun Dong, Song Wang, Yu Wang, Tyler Derr, Jundong Li
The low transparency on how the structure of the input network influences the bias in GNN outcome largely limits the safe adoption of GNNs in various decision-critical scenarios.
1 code implementation • 11 Oct 2023 • Osman Ülger, Yu Wang, Ysbrand Galama, Sezer Karaoglu, Theo Gevers, Martin R. Oswald
Humans have a remarkable ability to perceive and reason about the world around them by understanding the relationships between objects.
1 code implementation • bioRxiv 2023 • Tingpeng Yang, Tianze Ling, Boyan Sun, Zhendong Liang, Fan Xu, Xiansong Huang, Linhai Xie, Yonghong He, Leyuan Li, Fuchu He, Yu Wang, Cheng Chang
De novo peptide sequencing is a promising approach for novel peptide discovery.
1 code implementation • 2 Apr 2024 • Enshu Liu, Junyi Zhu, Zinan Lin, Xuefei Ning, Matthew B. Blaschko, Sergey Yekhanin, Shengen Yan, Guohao Dai, Huazhong Yang, Yu Wang
For example, LCSC achieves better performance using 1 number of function evaluation (NFE) than the base model with 2 NFE on consistency distillation, and decreases the NFE of DM from 15 to 9 while maintaining the generation quality on CIFAR-10.
1 code implementation • NeurIPS 2021 • Yu Wang, Jingyang Lin, Jingjing Zou, Yingwei Pan, Ting Yao, Tao Mei
Our work reveals a structured shortcoming of the existing mainstream self-supervised learning methods.
1 code implementation • 7 Dec 2022 • Yuying Zhao, Yu Wang, Tyler Derr
Although research efforts have been devoted to measuring and mitigating bias, they mainly study bias from the result-oriented perspective while neglecting the bias encoded in the decision-making procedure.
1 code implementation • 9 Feb 2023 • Dichao Liu, Toshihiko Yamasaki, Yu Wang, Kenji Mase, Jien Kato
Experimental results on the Statefarm Distracted Driver Detection Dataset and AUC Distracted Driver Dataset show that the proposed approach is highly effective for recognizing distracted driving behaviors from photos: (1) the teacher network's accuracy surpasses the previous best accuracy; (2) the student network achieves very high accuracy with only 0. 42M parameters (around 55% of the previous most lightweight model).
1 code implementation • CVPR 2023 • Tianli Zhang, Mengqi Xue, Jiangtao Zhang, Haofei Zhang, Yu Wang, Lechao Cheng, Jie Song, Mingli Song
Most existing online knowledge distillation(OKD) techniques typically require sophisticated modules to produce diverse knowledge for improving students' generalization ability.
1 code implementation • F1000Research 2023 • Xiaopeng Xu, Juexiao Zhou, Chen Zhu, Qing Zhan, Zhongxiao Li, Ruochi Zhang, Yu Wang, Xingyu Liao, Xin Gao
The SGPT-RL method achieved better results than Reinvent on the ACE2 task, where molecular docking was used as the optimization goal.
1 code implementation • 4 Mar 2024 • Yu Wang, Tongya Zheng, Yuxuan Liang, Shunyu Liu, Mingli Song
To address these challenges, we have tailored a Cross-city mObiLity trAnsformer (COLA) with a dedicated model-agnostic transfer framework by effectively transferring cross-city knowledge for human trajectory simulation.
1 code implementation • 6 Oct 2023 • Yu Wang, Tong Zhao, Yuying Zhao, Yunchao Liu, Xueqi Cheng, Neil Shah, Tyler Derr
Despite the widespread belief that low-degree nodes exhibit poorer LP performance, our empirical findings provide nuances to this viewpoint and prompt us to propose a better metric, Topological Concentration (TC), based on the intersection of the local subgraph of each node with the ones of its neighbors.
1 code implementation • 12 Dec 2023 • Zelin Xu, Tingsong Xiao, Wenchong He, Yu Wang, Zhe Jiang
The problem is challenging due to the sparse and noisy input labels, spatial uncertainty within the label inference process, and high computational costs associated with a large number of sample locations.
1 code implementation • 12 Jan 2024 • Pengfei Zhu, Qian Wang, Yu Wang, Jialu Li, QinGhua Hu
In this paper, we propose to dynamically learn the weights of SSL tasks for different nodes and fuse the embeddings learned from different SSL tasks to boost performance.
1 code implementation • 18 Jan 2024 • Zhenbang Xiao, Shunyu Liu, Yu Wang, Tongya Zheng, Mingli Song
Graph condensation has emerged as an intriguing technique to provide Graph Neural Networks for large-scale graphs with a more compact yet informative small graph to save the expensive costs of large-scale graph learning.
1 code implementation • 30 Jun 2022 • Huitong Chen, Yu Wang, QinGhua Hu
Re-balancing methods are used to alleviate the influence of data imbalance; however, we empirically discover that they would under-fit new classes.
1 code implementation • 1 Dec 2022 • Cheng Wang, Xue Fu, Yu Wang, Guan Gui, Haris Gacanin, Hikmet Sari, Fumiyuki Adachi
Specific emitter identification (SEI) is a potential physical layer authentication technology, which is one of the most critical complements of upper layer authentication.
1 code implementation • 2 Feb 2023 • Junbo Zhao, Xuefei Ning, Enshu Liu, Binxin Ru, Zixuan Zhou, Tianchen Zhao, Chen Chen, Jiajin Zhang, Qingmin Liao, Yu Wang
In the first step, we train different sub-predictors on different types of available low-fidelity information to extract beneficial knowledge as low-fidelity experts.
1 code implementation • 8 Feb 2023 • Xinyi Yang, Shiyu Huang, Yiwen Sun, Yuxiang Yang, Chao Yu, Wei-Wei Tu, Huazhong Yang, Yu Wang
Goal-conditioned hierarchical reinforcement learning (HRL) provides a promising direction to tackle this challenge by introducing a hierarchical structure to decompose the search space, where the low-level policy predicts primitive actions in the guidance of the goals derived from the high-level policy.
Hierarchical Reinforcement Learning Multi-agent Reinforcement Learning +2
1 code implementation • 15 Jan 2024 • Yuhao Wang, Yusheng Liao, Heyang Liu, Hongcheng Liu, Yu Wang, Yanfeng Wang
We believe that these hallucinations are partially due to the models' struggle with understanding what they can and cannot perceive from images, a capability we refer to as self-awareness in perception.
1 code implementation • 25 Apr 2019 • Han Xu, Junning Li, Liqiang Liu, Yu Wang, Haidong Yuan, Xin Wang
Measurement and estimation of parameters are essential for science and engineering, where one of the main quests is to find systematic schemes that can achieve high precision.
Quantum Physics Mesoscale and Nanoscale Physics
1 code implementation • 28 Aug 2021 • Yu Wang, Fang Liu, Daniele E. Schiavazzi
To reduce the computational cost without sacrificing inferential accuracy, we propose Normalizing Flow with Adaptive Surrogate (NoFAS), an optimization strategy that alternatively updates the normalizing flow parameters and surrogate model parameters.
1 code implementation • 8 Dec 2021 • Yu Wang, Alfred Hero
In this work, we study the emergence of sparsity and multiway structures in second-order statistical characterizations of dynamical processes governed by partial differential equations (PDEs).
1 code implementation • CVPR 2022 • Baisong Guo, Xiaoyun Zhang, HaoNing Wu, Yu Wang, Ya zhang, Yan-Feng Wang
Previous super-resolution (SR) approaches often formulate SR as a regression problem and pixel wise restoration, which leads to a blurry and unreal SR output.
1 code implementation • 17 Dec 2023 • Yu Wang, Zexue He, Zhankui He, Hao Xu, Julian McAuley
This fine-tuning allows the model to generate explanations that convey the compatibility relationships between items.
1 code implementation • 16 Jun 2017 • Bin Dai, Yu Wang, John Aston, Gang Hua, David Wipf
Variational autoencoders (VAE) represent a popular, flexible form of deep generative model that can be stochastically fit to samples from a given random process using an information-theoretic variational bound on the true underlying distribution.
1 code implementation • 7 Apr 2022 • Zeyu Sun, Monica G. Bobra, Xiantong Wang, Yu Wang, Hu Sun, Tamas Gombosi, Yang Chen, Alfred Hero
We consider the flare prediction problem that distinguishes flare-imminent active regions that produce an M- or X-class flare in the future 24 hours, from quiet active regions that do not produce any flare within $\pm 24$ hours.
1 code implementation • CVPR 2023 • Ying Ji, Yu Wang, Kensaku MORI, Jien Kato
Recent studies have achieved outstanding success in explaining 2D image recognition ConvNets.
1 code implementation • CVPR 2023 • Baowei Jiang, Bing Bai, Haozhe Lin, Yu Wang, Yuchen Guo, Lu Fang
Facial information is particularly sensitive in this regard.
1 code implementation • 10 Aug 2023 • Jiayuan Chen, Yu Wang, Ruining Deng, Quan Liu, Can Cui, Tianyuan Yao, Yilin Liu, Jianyong Zhong, Agnes B. Fogo, Haichun Yang, Shilin Zhao, Yuankai Huo
Podocytes, specialized epithelial cells that envelop the glomerular capillaries, play a pivotal role in maintaining renal health.
1 code implementation • 27 Dec 2023 • Yan Fan, Yu Wang, Pengfei Zhu, QinGhua Hu
In this work, we focus on semi-supervised continual learning (SSCL), where the model progressively learns from partially labeled data with unknown categories.
1 code implementation • 7 Jun 2019 • Xudong Sun, Alexej Gossmann, Yu Wang, Bernd Bischl
A novel variational inference based resampling framework is proposed to evaluate the robustness and generalization capability of deep learning models with respect to distribution shift.
1 code implementation • 1 Feb 2020 • Yu Wang, Byoungwook Jang, Alfred Hero
We apply the SyGlasso to an electroencephalography (EEG) study to compare the brain connectivity of alcoholic and nonalcoholic subjects.
1 code implementation • 12 May 2020 • Yu Wang, Rong Ge, Shuang Qiu
Unlike existing work in deep neural network (DNN) graphs optimization for inference performance, we explore DNN graph optimization for energy awareness and savings for power- and resource-constrained machine learning devices.
1 code implementation • 23 Nov 2021 • Tongya Zheng, Zunlei Feng, Yu Wang, Chengchao Shen, Mingli Song, Xingen Wang, Xinyu Wang, Chun Chen, Hao Xu
Our proposed Dynamic Preference Structure (DPS) framework consists of two stages: structure sampling and graph fusion.
1 code implementation • BMVC 2020 • Dichao Liu, Yu Wang, Jien Kato, Kenji Mase
The evaluation information is backpropagated and forces the classification stream to improve its awareness of visual attention, which helps classification.
Ranked #25 on Fine-Grained Image Classification on Stanford Cars
1 code implementation • 8 Jun 2023 • Shuo Ye, Shujian Yu, Wenjin Hou, Yu Wang, Xinge You
Fine-grained visual categorization (FGVC) is a challenging task due to similar visual appearances between various species.
1 code implementation • IEEE Transactions on Circuits and Systems for Video Technology 2023 • Guosong Jiang, Pengfei Zhu, Yu Wang, QinGhua Hu
In this paper, we point out that balancing between structural risk and open space risk is crucial for open set recognition, and re-formalize it as open set structural risk.
no code implementations • 18 Jun 2018 • Xuefei Ning, Yin Zheng, Zhuxi Jiang, Yu Wang, Huazhong Yang, Junzhou Huang
Moreover, we also propose HiTM-VAE, where the document-specific topic distributions are generated in a hierarchical manner.
no code implementations • 15 May 2018 • Xinghao Ding, Zhirui Lin, Fujin He, Yu Wang, Yue Huang
The estimation of crowd count in images has a wide range of applications such as video surveillance, traffic monitoring, public safety and urban planning.
no code implementations • 14 May 2018 • Wenshuo Li, Jincheng Yu, Xuefei Ning, Pengjun Wang, Qi Wei, Yu Wang, Huazhong Yang
So, in this paper, we propose a hardware-software collaborative attack framework to inject hidden neural network Trojans, which works as a back-door without requiring manipulating input images and is flexible for different scenarios.
no code implementations • 6 May 2018 • David Güera, Yu Wang, Luca Bondi, Paolo Bestagini, Stefano Tubaro, Edward J. Delp
We examine in this paper the problem of identifying the camera model or type that was used to take an image and that can be spoofed.
no code implementations • 8 Feb 2018 • Minhui Zou, Yang Shi, Chengliang Wang, Fangyu Li, WenZhan Song, Yu Wang
With the popularity of deep learning (DL), artificial intelligence (AI) has been applied in many areas of human life.
no code implementations • 1 Feb 2018 • Yu Wang, Xie Chen, Mark Gales, Anton Ragni, Jeremy Wong
As the combination approaches become more complicated the difference between the phonetic and graphemic systems further decreases.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
no code implementations • 1 Sep 2017 • Yu Wang, Jixing Xu, Aohan Wu, Mantian Li, Yang He, Jinghe Hu, Weipeng P. Yan
This paper proposes Telepath, a vision-based bionic recommender system model, which understands users from such perspective.
no code implementations • 18 Aug 2017 • Yu Wang, Jiayi Liu, Yuxiang Liu, Jun Hao, Yang He, Jinghe Hu, Weipeng P. Yan, Mantian Li
We present LADDER, the first deep reinforcement learning agent that can successfully learn control policies for large-scale real-world problems directly from raw inputs composed of high-level semantic information.
no code implementations • 18 Aug 2017 • Xie Chen, Xunying Liu, Anton Ragni, Yu Wang, Mark Gales
Instead of using a recurrent unit to capture the complete future word contexts, a feedforward unit is used to model a finite number of succeeding, future, words.
no code implementations • 16 Jun 2017 • Yuzhi Wang, Anqi Yang, Xiaoming Chen, Pengjun Wang, Yu Wang, Huazhong Yang
Temporal drift of sensory data is a severe problem impacting the data quality of wireless sensor networks (WSNs).
no code implementations • 24 May 2017 • Huizi Mao, Song Han, Jeff Pool, Wenshuo Li, Xingyu Liu, Yu Wang, William J. Dally
Since memory reference is more than two orders of magnitude more expensive than arithmetic operations, the regularity of sparse structure leads to more efficient hardware design.
no code implementations • 20 May 2017 • Yu Wang, Aniket Chakrabarti, David Sivakoff, Srinivasan Parthasarathy
In this work we devise an effective and efficient three-step-approach for detecting change points in dynamic networks under the snapshot model.
no code implementations • 1 Dec 2016 • Song Han, Junlong Kang, Huizi Mao, Yiming Hu, Xin Li, Yubin Li, Dongliang Xie, Hong Luo, Song Yao, Yu Wang, Huazhong Yang, William J. Dally
Evaluated on the LSTM for speech recognition benchmark, ESE is 43x and 3x faster than Core i7 5930k CPU and Pascal Titan X GPU implementations.
no code implementations • 6 Oct 2016 • Yu Wang, Haofu Liao, Yang Feng, Xiangyang Xu, Jiebo Luo
We find that Chinese, Japanese and Koreans do exhibit substantial differences in certain attributes, such as bangs, smiling, and bushy eyebrows.
no code implementations • 16 Oct 2015 • Kumpati S. Narendra, Snehasis Mukhopadyhay, Yu Wang
During the past two decades, the first author has worked on numerous methods for improving the stability, robustness, and performance of adaptive systems using multiple models and the other authors have collaborated with him on some of them.
no code implementations • 31 Mar 2014 • Xiangyu Chang, Yu Wang, Rongjian Li, Zongben Xu
Nevertheless, this framework has two serious drawbacks: One is that the solution of the framework unavoidably involves a considerable portion of redundant noise features in many situations, and the other is that the framework neither offers intuitive explanations on why this framework can select relevant features nor leads to any theoretical guarantee for feature selection consistency.
no code implementations • 24 Dec 2017 • Kaiyuan Guo, Shulin Zeng, Jincheng Yu, Yu Wang, Huazhong Yang
Various FPGA based accelerator designs have been proposed with software and hardware optimization techniques to achieve high speed and energy efficiency.
Hardware Architecture
no code implementations • 18 Sep 2018 • Yilin Shen, Xiangyu Zeng, Yu Wang, Hongxia Jin
The results show that our approach leverages such simple user information to outperform state-of-the-art approaches by 0. 25% for intent detection and 0. 31% for slot filling using standard training data.
no code implementations • 30 Oct 2018 • Anton Ragni, Qiujia Li, Mark Gales, Yu Wang
These errors are not accounted for by the standard confidence estimation schemes and are hard to rectify in the upstream and downstream processing.
no code implementations • COLING 2018 • Yu Wang, Abhishek Patel, Hongxia Jin
In this paper, a new deep reinforcement learning based augmented general tagging system is proposed.
no code implementations • ICLR 2019 • Yu Wang, Jack W. Stokes, Mady Marinescu
Antimalware products are a key component in detecting malware attacks, and their engines typically execute unknown programs in a sandbox prior to running them on the native operating system.
no code implementations • ICLR 2019 • Yu Wang, Fengjuan Gao, Amin Alipour, Linzhang Wang, Xuandong Li, Zhendong Su
Boolean satisfiability (SAT) is one of the most well-known NP-complete problems and has been extensively studied.
no code implementations • ICLR 2018 • Xuefei Ning, Yin Zheng, Zhuxi Jiang, Yu Wang, Huazhong Yang, Junzhou Huang
On the other hand, different with the other BNP topic models, the inference of iTM-VAE is modeled by neural networks, which has rich representation capacity and can be computed in a simple feed-forward manner.
no code implementations • 26 Dec 2018 • Yu Wang, Abhishek Patel, Hongxia Jin
In this paper, a new deep reinforcement learning based augmented general sequence tagging system is proposed.
no code implementations • 22 Feb 2019 • Yu Wang, Siqi Wu, Bin Yu
First, we obtain a necessary and sufficient norm condition for the reference dictionary $D^*$ to be a sharp local minimum of the expected $\ell_1$ objective function.
no code implementations • CVPR 2019 • Yingwei Pan, Ting Yao, Yehao Li, Yu Wang, Chong-Wah Ngo, Tao Mei
Specifically, we present Transferrable Prototypical Networks (TPN) for adaptation such that the prototypes for each class in source and target domains are close in the embedding space and the score distributions predicted by prototypes separately on source and target data are similar.
no code implementations • 19 Jul 2019 • Shuqiang Lu, Lingyun Ying, Wenjie Lin, Yu Wang, Meining Nie, Kaiwen Shen, Lu Liu, Haixin Duan
With the development of artificial intelligence algorithms like deep learning models and the successful applications in many different fields, further similar trails of deep learning technology have been made in cyber security area.
no code implementations • IJCNLP 2019 • Yu Wang
In this paper, we present a fast and reliable method based on PCA to select the number of dimensions for word embeddings.
no code implementations • 30 Sep 2019 • Lin-Lin Wang, Yu Wang, Mark J. F. Gales
These systems are explored for non-native spoken English data in this paper.
no code implementations • 2 Oct 2019 • Hao Zhou, Jorge Laval, Anye Zhou, Yu Wang, Wenchao Wu, Zhu Qing, Srinivas Peeta
Some suggestions towards congestion mitigation for future mMP studies are proposed: i) enrich data collection to facilitate the congestion learning, ii) incorporate non-imitation learning methods to combine traffic efficiency into a safety-oriented technical route, and iii) integrate domain knowledge from the traditional car following (CF) theory to improve the string stability of mMP.
no code implementations • 9 Feb 2020 • Yu Wang, Yining Sun, Zuchang Ma, Lisheng Gao, Yang Xu, Ting Sun
Then, we apply these pre-training models to a NER task by fine-tuning, and compare the effects of the different model architecture and pre-training tasks on the NER task.
no code implementations • 21 Mar 2020 • Yang Feng, Yu Wang, Jiebo Luo
In this paper, we introduce a novel gating mechanism to deep neural networks.
Optical Flow Estimation Video-Based Person Re-Identification
no code implementations • 20 Mar 2020 • Xuefei Ning, Guangjun Ge, Wenshuo Li, Zhenhua Zhu, Yin Zheng, Xiaoming Chen, Zhen Gao, Yu Wang, Huazhong Yang
By inspecting the discovered architectures, we find that the operation primitives, the weight quantization range, the capacity of the model, and the connection pattern have influences on the fault resilience capability of NN models.
no code implementations • 26 Mar 2020 • Shulin Zeng, Guohao Dai, Hanbo Sun, Kai Zhong, Guangjun Ge, Kaiyuan Guo, Yu Wang, Huazhong Yang
Currently, the majority of FPGA-based DNN accelerators in the cloud run in a time-division multiplexing way for multiple users sharing a single FPGA, and require re-compilation with $\sim$100 s overhead.
no code implementations • 1 Apr 2020 • Yu Wang, Hussein Sibai, Mark Yen, Sayan Mitra, Geir E. Dullerud
We also show that the standard exponential mechanism that randomizes the output of an algorithm to achieve differential privacy fails to do so in the context of sequential algorithms.
no code implementations • 1 Apr 2020 • Yu Wang, Nima Roohi, Matthew West, Mahesh Viswanathan, Geir E. Dullerud
Probabilistic Computation Tree Logic (PCTL) is frequently used to formally specify control objectives such as probabilistic reachability and safety.
no code implementations • 21 Apr 2020 • Viet Duong, Phu Pham, Tongyu Yang, Yu Wang, Jiebo Luo
Recently, the pandemic of the novel Coronavirus Disease-2019 (COVID-19) has presented governments with ultimate challenges.
no code implementations • 21 Apr 2020 • Long Chen, Hanjia Lyu, Tongyu Yang, Yu Wang, Jiebo Luo
To model the substantive difference of tweets with controversial terms and those with non-controversial terms, we apply topic modeling and LIWC-based sentiment analysis.
no code implementations • 8 Nov 2019 • Xiaoming Chen, Yinhe Han, Yu Wang
Evaluations based on the 65nm technology demonstrate that the proposed architecture nearly reaches the theoretical minimum communication in a three-level memory hierarchy and it is computation dominant.
Distributed, Parallel, and Cluster Computing Hardware Architecture
no code implementations • 2 May 2020 • Yu Wang, Yuelin Wang, Jie Liu, Zhuo Liu
More importantly, we discuss four kinds of basic approaches, including statistical machine translation based approach, neural machine translation based approach, classification based approach and language model based approach, six commonly applied performance boosting techniques for GEC systems and two data augmentation methods.
no code implementations • 18 May 2020 • Yu Wang, Fengjuan Gao, Linzhang Wang, Ke Wang
We have also created a neural bug detector based on GINN to catch null pointer deference bugs in Java code.
no code implementations • 22 May 2020 • Kechen Qin, Yu Wang, Cheng Li, Kalpa Gunaratna, Hongxia Jin, Virgil Pavlu, Javed A. Aslam
Multi-hop knowledge based question answering (KBQA) is a complex task for natural language understanding.
no code implementations • 27 May 2020 • Yu Wang, Junpeng Bao, JianQiang Du, Yongfeng Li
Compared with the existing AKI predictors, the predictor in this work greatly improves the precision of early prediction of AKI by using the Convolutional Neural Network architecture and a more concise input vector.
no code implementations • 4 Jun 2020 • Kai Zhong, Xuefei Ning, Guohao Dai, Zhenhua Zhu, Tianchen Zhao, Shulin Zeng, Yu Wang, Huazhong Yang
For training a variety of models on CIFAR-10, using 1-bit mantissa and 2-bit exponent is adequate to keep the accuracy loss within $1\%$.
no code implementations • 11 Jun 2020 • Yu Wang, Qitong Gao, Miroslav Pajic
For monotonicity constraints, we propose to use nonnegative neural networks and batch normalization.
no code implementations • 28 Jun 2020 • Sijia Chen, Yu Wang, Li Huang, Runzhou Ge, Yihan Hu, Zhuangzhuang Ding, Jie Liao
A practical autonomous driving system urges the need to reliably and accurately detect vehicles and persons.
no code implementations • 28 Jun 2020 • Yu Wang, Sijia Chen, Li Huang, Runzhou Ge, Yihan Hu, Zhuangzhuang Ding, Jie Liao
This technical report presents the online and real-time 2D and 3D multi-object tracking (MOT) algorithms that reached the 1st places on both Waymo Open Dataset 2D tracking and 3D tracking challenges.
no code implementations • 28 Jun 2020 • Zhuangzhuang Ding, Yihan Hu, Runzhou Ge, Li Huang, Sijia Chen, Yu Wang, Jie Liao
We proposed a one-stage, anchor-free and NMS-free 3D point cloud object detector AFDet, using object key-points to encode the 3D attributes, and to learn an end-to-end point cloud object detection without the need of hand-engineering or learning the anchors.
no code implementations • 1 Jul 2020 • Yi-Peng Zhang, Hanjia Lyu, Yubao Liu, Xiyang Zhang, Yu Wang, Jiebo Luo
The COVID-19 pandemic has severely affected people's daily lives and caused tremendous economic loss worldwide.
no code implementations • 13 Jul 2020 • Yucan Zhou, Yu Wang, Jianfei Cai, Yu Zhou, QinGhua Hu, Weiping Wang
Some works in the optimization of deep neural networks have shown that a better arrangement of training data can make the classifier converge faster and perform better.
no code implementations • 31 Jul 2020 • Tong Wu, Xuefei Ning, Wenshuo Li, Ranran Huang, Huazhong Yang, Yu Wang
In this paper, we tackle the issue of physical adversarial examples for object detectors in the wild.
no code implementations • 13 Aug 2020 • Marco Manfredi, Yu Wang
Robustness to small image translations is a highly desirable property for object detectors.
no code implementations • 11 Sep 2020 • Mark Cartwright, Jason Cramer, Ana Elisa Mendez Mendez, Yu Wang, Ho-Hsiang Wu, Vincent Lostanlen, Magdalena Fuentes, Graham Dove, Charlie Mydlarz, Justin Salamon, Oded Nov, Juan Pablo Bello
In this article, we describe our data collection procedure and propose evaluation metrics for multilabel classification of urban sound tags.
no code implementations • 13 Sep 2020 • Zishen Wan, Bo Yu, Thomas Yuang Li, Jie Tang, Yuhao Zhu, Yu Wang, Arijit Raychowdhury, Shaoshan Liu
On the other hand, FPGA-based robotic accelerators are becoming increasingly competitive alternatives, especially in latency-critical and power-limited scenarios.
no code implementations • 1 Jan 2021 • Chao Yu, Akash Velu, Eugene Vinitsky, Yu Wang, Alexandre Bayen, Yi Wu
We benchmark commonly used multi-agent deep reinforcement learning (MARL) algorithms on a variety of cooperative multi-agent games.
no code implementations • 1 Jan 2021 • Kai Zhong, Xuefei Ning, Tianchen Zhao, Zhenhua Zhu, Shulin Zeng, Guohao Dai, Yu Wang, Huazhong Yang
Through this dynamic precision framework, we can reduce the bit-width of convolution, which is the most computational cost, while keeping the training process close to the full precision floating-point training.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Shayne Longpre, Yu Wang, Christopher DuBois
Task-agnostic forms of data augmentation have proven widely effective in computer vision, even on pretrained models.