no code implementations • 2 Jun 2018 • Pinlong Zhao, Zhouyu Fu, Ou wu, QinGhua Hu, Jun Wang
In contrast to existing defense methods, the proposed method does not require knowledge of the process for generating adversarial examples and can be applied to defend against different types of attacks.
no code implementations • 5 Jul 2017 • Haifeng Zhang, Jun Wang, Zhiming Zhou, Wei-Nan Zhang, Ying Wen, Yong Yu, Wenxin Li
In typical reinforcement learning (RL), the environment is assumed given and the goal of the learning is to identify an optimal policy for the agent taking actions through its interactions with the environment.
no code implementations • 17 May 2018 • Jun Wang, Sujoy Sikdar, Tyler Shepherd, Zhibing Zhao, Chunheng Jiang, Lirong Xia
We also propose novel ILP formulations for PUT-winners under STV and RP, respectively.
no code implementations • 13 Sep 2017 • Yaodong Yang, Lantao Yu, Yiwei Bai, Jun Wang, Wei-Nan Zhang, Ying Wen, Yong Yu
We conduct an empirical study on discovering the ordered collective dynamics obtained by a population of intelligence agents, driven by million-agent reinforcement learning.
no code implementations • 21 Apr 2018 • Jihua Zhu, Siyu Xu, Zutao Jiang, Shanmin Pang, Jun Wang, Zhongyu Li
This paper proposes a global approach for the multi-view registration of unordered range scans.
no code implementations • 25 Sep 2017 • Congcong Jin, Jihua Zhu, Yaochen Li, Shanmin Pang, Lei Chen, Jun Wang
Then, it proposes the weighted LRS decomposition, where each block element is assigned with one estimated weight to denote its reliability.
no code implementations • 6 Mar 2018 • Jingwei Song, Jun Wang, Liang Zhao, Shoudong Huang, Gamini Dissanayake
Idled CPU is used to perform ORB- SLAM for providing robust global pose.
no code implementations • 15 Mar 2018 • Sidi Lu, Yaoming Zhu, Wei-Nan Zhang, Jun Wang, Yong Yu
This paper presents a systematic survey on recent development of neural text generation models.
no code implementations • 1 Mar 2018 • Kan Ren, Wei-Nan Zhang, Ke Chang, Yifei Rong, Yong Yu, Jun Wang
From the learning perspective, we show that the bidding machine can be updated smoothly with both offline periodical batch or online sequential training schemes.
no code implementations • 7 Mar 2018 • Bowen Wu, Zhangling Chen, Jun Wang, Huaming Wu
With the remarkable success achieved by the Convolutional Neural Networks (CNNs) in object recognition recently, deep learning is being widely used in the computer vision community.
no code implementations • 5 Aug 2017 • Zhiming Zhou, Wei-Nan Zhang, Jun Wang
In this article, we mathematically study several GAN related topics, including Inception score, label smoothing, gradient vanishing and the -log(D(x)) alternative.
no code implementations • 27 Feb 2018 • Junqi Jin, Chengru Song, Han Li, Kun Gai, Jun Wang, Wei-Nan Zhang
Real-time advertising allows advertisers to bid for each impression for a visiting user.
no code implementations • 25 Feb 2015 • Yu-Gang Jiang, Zuxuan Wu, Jun Wang, xiangyang xue, Shih-Fu Chang
In this paper, we study the challenging problem of categorizing videos according to high-level semantics such as the existence of a particular human action or a complex event.
no code implementations • 15 Jan 2018 • Ao Zhang, Nan Li, Jian Pu, Jun Wang, Junchi Yan, Hongyuan Zha
Learning a classifier with control on the false-positive rate plays a critical role in many machine learning applications.
no code implementations • 31 Dec 2017 • Simon Stiebellehner, Jun Wang, Shuai Yuan
In order to maximize the predictive performance of our look-alike modeling algorithms, we propose two novel hybrid filtering techniques that utilize the recent neural probabilistic language model algorithm doc2vec.
no code implementations • 29 Nov 2017 • Rafał Muszyński, Jun Wang
We find that the agents that achieve higher happiness during testing against hand-coded AI, have lower happiness when competing against each other.
no code implementations • 30 Nov 2017 • Rui Luo, Wei-Nan Zhang, Xiaojun Xu, Jun Wang
In this paper, we show that the recent integration of statistical models with deep recurrent neural networks provides a new way of formulating volatility (the degree of variation of time series) models that have been widely used in time series analysis and prediction in finance.
no code implementations • 2 Nov 2017 • I-Hong Jhuo, Jun Wang
In this paper, we consider the fundamental problem of finding a nearest set from a collection of sets, to a query set.
no code implementations • 28 Apr 2017 • Minmin Xu, Siyu Xu, Jihua Zhu, Yaochen Li, Jun Wang, Huimin Lu
This paper proposes an effective approach for the scaling registration of $m$-D point sets.
no code implementations • 2 Dec 2016 • Nurjahan Begum, Liudmila Ulanova, Hoang Anh Dau, Jun Wang, Eamonn Keogh
Clustering time series under DTW remains a computationally expensive operation.
no code implementations • 22 Jun 2016 • Ying Wen, Wei-Nan Zhang, Rui Luo, Jun Wang
Recently, the rapid development of word embedding and neural networks has brought new inspiration to various NLP and IR tasks.
no code implementations • 19 Sep 2014 • Zhaohong Deng, Kup-Sze Choi, Yizhang Jiang, Jun Wang, Shitong Wang
Subspace clustering (SC) is a promising clustering technology to identify clusters based on their associations with subspaces in high dimensional spaces.
no code implementations • 16 Mar 2016 • Kleanthis Malialis, Jun Wang, Gary Brooks, George Frangou
In this paper, we formulate feature selection as a multiagent coordination problem and propose a novel feature selection method using multiagent reinforcement learning.
no code implementations • 11 Jan 2016 • Wei-Nan Zhang, Lingxi Chen, Jun Wang
In this work, we propose a general framework which learns the user profiles based on their online browsing behaviour, and transfers the learned knowledge onto prediction of their ad response.
no code implementations • 21 Sep 2015 • Zuxuan Wu, Yu-Gang Jiang, Xi Wang, Hao Ye, xiangyang xue, Jun Wang
A multi-stream framework is proposed to fully utilize the rich multimodal information in videos.
no code implementations • 4 Nov 2015 • Phong Nguyen, Jun Wang, Alexandros Kalousis
Motivated by the fact that very often the users' and items' descriptions as well as the preference behavior can be well summarized by a small number of hidden factors, we propose a novel algorithm, LambdaMART Matrix Factorization (LambdaMART-MF), that learns a low rank latent representation of users and items using gradient boosted trees.
no code implementations • 17 Sep 2015 • Jun Wang, Wei Liu, Sanjiv Kumar, Shih-Fu Chang
Such learning to hash methods exploit information such as data distributions or class labels when optimizing the hash codes or functions.
no code implementations • 8 Sep 2015 • Jianwei Luo, Jianguo Li, Jun Wang, Zhiguo Jiang, Yurong Chen
Results show that deep attribute approaches achieve state-of-the-art results, and outperforms existing peer methods with a significant margin, even though some benchmarks have little overlap of concepts with the pre-trained CNN models.
no code implementations • 12 May 2014 • Jun Wang, Ke Sun, Fei Sha, Stephane Marchand-Maillet, Alexandros Kalousis
This induces in the input data space a new family of distance metric with unique properties.
no code implementations • 26 Apr 2013 • Leon Derczynski, Richard Shaw, Ben Solway, Jun Wang
Question answering involves developing methods to extract useful information from large collections of documents.
no code implementations • 20 May 2014 • Bo-Wei Chen, Shuai Yuan, Jun Wang
From the experiments we find that, in a less competitive market, lower prices of the guaranteed contracts will encourage the purchase in advance and the revenue gain is mainly contributed by the increased competition in future RTB.
Computer Science and Game Theory
no code implementations • 10 Sep 2018 • Weixun Wang, Junqi Jin, Jianye Hao, Chunjie Chen, Chuan Yu, Wei-Nan Zhang, Jun Wang, Xiaotian Hao, Yixi Wang, Han Li, Jian Xu, Kun Gai
In this paper, we investigate the problem of advertising with adaptive exposure: can we dynamically determine the number and positions of ads for each user visit under certain business constraints so that the platform revenue can be increased?
no code implementations • 10 Oct 2018 • Zheng Tian, Shihao Zou, Ian Davies, Tim Warr, Lisheng Wu, Haitham Bou Ammar, Jun Wang
The auxiliary reward for communication is integrated into the learning of the policy module.
no code implementations • 5 Nov 2018 • Lisheng Wu, Minne Li, Jun Wang
Humans have consciousness as the ability to perceive events and objects: a mental model of the world developed from the most impoverished of visual stimuli, enabling humans to make rapid decisions and take actions.
no code implementations • 14 Nov 2018 • Haifeng Zhang, Zilong Guo, Han Cai, Chris Wang, Wei-Nan Zhang, Yong Yu, Wenxin Li, Jun Wang
With the rapid growth of the express industry, intelligent warehouses that employ autonomous robots for carrying parcels have been widely used to handle the vast express volume.
no code implementations • 15 Dec 2018 • Guanghua Pan, Jun Wang, Rendong Ying, Peilin Liu
Deep learning on point clouds has made a lot of progress recently.
no code implementations • NeurIPS 2015 • Ke Sun, Jun Wang, Alexandros Kalousis, Stephane Marchand-Maillet
We give theoretical propositions to show that space-time is a more powerful representation than Euclidean space.
no code implementations • NeurIPS 2012 • Jun Wang, Alexandros Kalousis, Adam Woznica
We present a new parametric local metric learning method in which we learn a smooth metric matrix function over the data manifold.
no code implementations • NeurIPS 2011 • Jun Wang, Huyen T. Do, Adam Woznica, Alexandros Kalousis
However, the problem then becomes finding the appropriate kernel function.
no code implementations • 9 Jan 2019 • Peng Xu, Zhaohong Deng, Jun Wang, Qun Zhang, Shitong Wang
A core issue in transfer learning is to learn a shared feature space in where the distributions of the data from two domains are matched.
no code implementations • 14 Jan 2019 • Yourui Huangfu, Jian Wang, Rong Li, Chen Xu, Xianbin Wang, Huazi Zhang, Jun Wang
Accurate prediction of fading channel in future is essential to realize adaptive transmission and other methods that can save power and provide gains.
no code implementations • 26 Jan 2019 • Ying Wen, Yaodong Yang, Rui Luo, Jun Wang
Though limited in real-world decision making, most multi-agent reinforcement learning (MARL) models assume perfectly rational agents -- a property hardly met due to individual's cognitive limitation and/or the tractability of the decision problem.
no code implementations • ICLR 2019 • Ying Wen, Yaodong Yang, Rui Luo, Jun Wang, Wei Pan
Our methods are tested on both the matrix game and the differential game, which have a non-trivial equilibrium where common gradient-based methods fail to converge.
Multi-agent Reinforcement Learning reinforcement-learning +1
no code implementations • 16 Jan 2019 • Jun Wang, Sujoy Sikdar, Tyler Shepherd, Zhibing Zhao, Chunheng Jiang, Lirong Xia
STV and ranked pairs (RP) are two well-studied voting rules for group decision-making.
no code implementations • CVPR 2016 • Hao Wang, Jun Wang, Wang Liang
A system capable of performing robust online volumetric reconstruction of indoor scenes based on input from a handheld RGB-D camera is presented.
no code implementations • 4 Feb 2019 • Dongyan Guo, Jun Wang, Weixuan Zhao, Ying Cui, Zhenhua Wang, Sheng-Yong Chen
Both features and the channel weights are utilized in a template generation layer to generate a discriminative template.
no code implementations • 11 Sep 2018 • Yong Chen, Ming Zhou, Ying Wen, Yaodong Yang, Yufeng Su, Wei-Nan Zhang, Dell Zhang, Jun Wang, Han Liu
Deep Q-learning has achieved a significant success in single-agent decision making tasks.
Multiagent Systems
no code implementations • 22 Feb 2019 • Xianbin Wang, Huazi Zhang, Rong Li, Lingchen Huang, Shengchen Dai, Yourui Huangfu, Jun Wang
Specifically, before each SC decoding attempt, a long short-term memory (LSTM) network is exploited to either (i) locate the first error bit, or (ii) undo a previous `wrong' flip.
no code implementations • 4 Mar 2019 • Minne Li, Zheng Tian, Pranav Nashikkar, Ian Davies, Ying Wen, Jun Wang
Existing model-based reinforcement learning methods often study perception modeling and decision making separately.
no code implementations • 16 Apr 2019 • Lingchen Huang, Huazi Zhang, Rong Li, Yiqun Ge, Jun Wang
In this paper, we model nested polar code construction as a Markov decision process (MDP), and tackle it with advanced reinforcement learning (RL) techniques.
no code implementations • 24 Apr 2019 • Peng Xu, Zhaohong Deng, Chen Cui, Te Zhang, Kup-Sze Choi, Gu Suhang, Jun Wang, Shitong Wang
Furthermore, for highly nonlinear modeling task, it is usually necessary to use a large number of rules which further weakens the clarity and interpretability of TSK FS.
no code implementations • 10 May 2019 • Xing Wang, Jun Wang, Carlotta Domeniconi, Guoxian Yu, Guo-Qiang Xiao, Maozu Guo
To ease this process, we consider diverse clusterings embedded in different subspaces, and analyze the embedding subspaces to shed light into the structure of each clustering.
no code implementations • 13 May 2019 • Yuying Xing, Guoxian Yu, Carlotta Domeniconi, Jun Wang, Zili Zhang, Maozu Guo
To preserve the intrinsic structure of the data matrices, M3Lcmf collaboratively factorizes them into low-rank matrices, explores the latent relationships between bags, instances, and labels, and selectively merges the data matrices.
no code implementations • 13 May 2019 • Shixing Yao, Guoxian Yu, Jun Wang, Carlotta Domeniconi, Xiangliang Zhang
It then uses matrix factorization on the individual matrices, along with the shared matrix, to generate diverse clusterings of high-quality.
no code implementations • 13 May 2019 • Hongpeng Zhou, Minghao Yang, Jun Wang, Wei Pan
One-Shot Neural Architecture Search (NAS) is a promising method to significantly reduce search time without any separate training.
no code implementations • 11 May 2019 • Xuanwu Liu, Guoxian Yu, Carlotta Domeniconi, Jun Wang, Yazhou Ren, Maozu Guo
Next, to expand the semantic representation power of hand-crafted features, RDCMH integrates the semantic ranking information into deep cross-modal hashing and jointly optimizes the compatible parameters of deep feature representations and of hashing functions.
no code implementations • 14 May 2019 • Xia Chen, Guoxian Yu, Jun Wang, Carlotta Domeniconi, Zhao Li, Xiangliang Zhang
To maximize the profit of utilizing the rare and valuable supervised information in HNEs, we develop a novel Active Heterogeneous Network Embedding (ActiveHNE) framework, which includes two components: Discriminative Heterogeneous Network Embedding (DHNE) and Active Query in Heterogeneous Networks (AQHN).
no code implementations • 15 May 2019 • Jian Wang, Chen Xu, Yourui Huangfu, Rong Li, Yiqun Ge, Jun Wang
Integrating artificial intelligence (AI) into wireless networks has drawn significant interest in both industry and academia.
no code implementations • 16 May 2019 • Jun Wang, Dan Su, Jie Chen, Shulin Feng, Dongpeng Ma, Na Li, Dong Yu
We propose a novel method which simultaneously models both the sequence discriminative training and the feature discriminative learning within a single network architecture, so that it can learn discriminative deep features in sequence training that obviates the need for presegmented training data.
no code implementations • 22 May 2019 • Peng Xu, Zhaohong Deng, Kup-Sze Choi, Jun Wang, Shitong Wang
The two domains often lie in different feature spaces due to diverse data collection methods, which leads to the more challenging task of heterogeneous domain adaptation (HDA).
no code implementations • 29 May 2019 • Rui Luo, Qiang Zhang, Yaodong Yang, Jun Wang
In this paper, we present a new practical method for Bayesian learning that can rapidly draw representative samples from complex posterior distributions with multiple isolated modes in the presence of mini-batch noise.
no code implementations • 29 May 2019 • Xuanwu Liu, Jun Wang, Guoxian Yu, Carlotta Domeniconi, Xiangliang Zhang
FlexCMH first introduces a clustering-based matching strategy to explore the local structure of each cluster, and thus to find the potential correspondence between clusters (and samples therein) across modalities.
no code implementations • 27 May 2019 • Jiarui Jin, Ming Zhou, Wei-Nan Zhang, Minne Li, Zilong Guo, Zhiwei Qin, Yan Jiao, Xiaocheng Tang, Chenxi Wang, Jun Wang, Guobin Wu, Jieping Ye
How to optimally dispatch orders to vehicles and how to trade off between immediate and future returns are fundamental questions for a typical ride-hailing platform.
Multiagent Systems
no code implementations • 11 May 2019 • Dong Li, Qichao Zhang, Dongbin Zhao, Yuzheng Zhuang, Bin Wang, Wulong Liu, Rasul Tutunov, Jun Wang
To address the long-term memory issue, this paper proposes a graph attention memory (GAM) architecture consisting of memory construction module, graph attention module and control module.
no code implementations • WS 2019 • Beiming Cao, Nordine Sebkhi, Ted Mau, Omer T. Inan, Jun Wang
Articulation-to-speech (ATS) synthesis is a software design in SSI that directly converts articulatory movement information into audible speech signals.
no code implementations • WS 2019 • Alan Wisler, Kristin Teplansky, Jordan Green, Yana Yunusova, Thomas Campbell, Daragh Heitzman, Jun Wang
Experimental results demonstrated the AFSFRS-R bulbar subscore can be predicted from speech samples, which has clinical implication for automatic monitoring of the disease progression of ALS using speech information.
no code implementations • 22 Jul 2019 • Yourui Huangfu, Jian Wang, Chen Xu, Rong Li, Yiqun Ge, Xianbin Wang, Huazi Zhang, Jun Wang
In this paper, we propose a neural-network-based realistic channel model with both the similar accuracy as deterministic channel models and uniformity as stochastic channel models.
no code implementations • 30 Jul 2019 • Mohammed Amin Abdullah, Hang Ren, Haitham Bou Ammar, Vladimir Milenkovic, Rui Luo, Mingtian Zhang, Jun Wang
Reinforcement learning algorithms, though successful, tend to over-fit to training environments hampering their application to the real-world.
no code implementations • 19 Aug 2019 • Dagui Chen, Junqi Jin, Wei-Nan Zhang, Fei Pan, Lvyin Niu, Chuan Yu, Jun Wang, Han Li, Jian Xu, Kun Gai
We refer to this process as Leverage.
no code implementations • 19 Aug 2019 • Xuanwu Liu, Zhao Li, Jun Wang, Guoxian Yu, Carlotta Domeniconi, Xiangliang Zhang
It then defines an objective function to achieve deep feature learning compatible with the composite similarity preserving, category attribute space learning, and hashing coding function learning.
no code implementations • 3 Sep 2019 • Vasco Lopes, Fabio Maria Carlucci, Pedro M Esperança, Marco Singh, Victor Gabillon, Antoine Yang, Hang Xu, Zewei Chen, Jun Wang
The Neural Architecture Search (NAS) problem is typically formulated as a graph search problem where the goal is to learn the optimal operations over edges in order to maximise a graph-level global objective.
no code implementations • 15 Sep 2019 • Shihao Zou, Zhonghua Li, Mohammad Akbari, Jun Wang, Peng Zhang
By defining reward as a function of NDCG, we can optimize our model directly on the ranking performance measure.
no code implementations • 11 Oct 2019 • Tianshuo Zhou, Ziyang Li, Gong Cheng, Jun Wang, Yu'Ang Wei
The model applies to meta-path based relevance where a meta-path characterizes a particular type of semantics of relating the query entity to answer entities.
no code implementations • 28 Oct 2019 • Max W. Y. Lam, Jun Wang, Dan Su, Dong Yu
Deep-learning based speech separation models confront poor generalization problem that even the state-of-the-art models could abruptly fail when evaluating them in mismatch conditions.
no code implementations • IJCNLP 2019 • Bei Yu, Yingya Li, Jun Wang
We then applied the prediction model to measure the causal language use in the research conclusions of about 38, 000 observational studies in PubMed.
no code implementations • 7 Nov 2019 • Jinzheng Tu, Guoxian Yu, Carlotta Domeniconi, Jun Wang, Xiangliang Zhang
AMCC accounts for the commonality and individuality of workers, and assumes that workers can be organized into different groups.
no code implementations • 13 Nov 2019 • Chen Xu, Jian Wang, Tianhang Yu, Chuili Kong, Yourui Huangfu, Rong Li, Yiqun Ge, Jun Wang
In this paper, the downlink packet scheduling problem for cellular networks is modeled, which jointly optimizes throughput, fairness and packet drop rate.
no code implementations • 26 Nov 2019 • Shaowei Wei, Jun Wang, Guoxian Yu, Carlotta, Xiangliang Zhang
Multi-view clustering aims at integrating complementary information from multiple heterogeneous views to improve clustering results.
no code implementations • 3 Dec 2019 • Hangyu Mao, Wulong Liu, Jianye Hao, Jun Luo, Dong Li, Zhengchao Zhang, Jun Wang, Zhen Xiao
Social psychology and real experiences show that cognitive consistency plays an important role to keep human society in order: if people have a more consistent cognition about their environments, they are more likely to achieve better cooperation.
no code implementations • 24 Dec 2019 • Jingzheng Tu, Guoxian Yu, Jun Wang, Carlotta Domeniconi, Xiangliang Zhang
However, they all assume that workers' label quality is stable over time (always at the same level whenever they conduct the tasks).
no code implementations • 27 Dec 2019 • Jun Wang, Hefu Zhang, Qi Liu, Zhen Pan, Hanqing Tao
However, few of them take into account the inherent decision-making process between investors and crowdfunding dynamics.
no code implementations • 10 Feb 2020 • Rasul Tutunov, Minne Li, Alexander I. Cowen-Rivers, Jun Wang, Haitham Bou-Ammar
In this paper, we present C-ADAM, the first adaptive solver for compositional problems involving a non-linear functional nesting of expected values.
no code implementations • 11 Feb 2020 • Junjie Sheng, Xiangfeng Wang, Bo Jin, Junchi Yan, Wenhao Li, Tsung-Hui Chang, Jun Wang, Hongyuan Zha
This work explores the large-scale multi-agent communication mechanism under a multi-agent reinforcement learning (MARL) setting.
Multi-agent Reinforcement Learning reinforcement-learning +1
no code implementations • 3 Mar 2020 • Ziling Wu, Ping Liu, Zheng Hu, Bocheng Li, Jun Wang
Our methods can significantly reduce the cost of development and maintenance of anomaly detection.
no code implementations • 17 Mar 2020 • Tingting Yu, Guoxian Yu, Jun Wang, Maozu Guo
Partial multi-label learning (PML) models the scenario where each training instance is annotated with a set of candidate labels, and only some of the labels are relevant.
no code implementations • 22 Mar 2020 • Jingwei Song, Shaobo Xia, Jun Wang, Dong Chen
To this end, we propose a new framework for curved building reconstruction via assembling and deforming geometric primitives.
no code implementations • 22 Mar 2020 • Jingwei Song, Jun Wang, Liang Zhao, Shoudong Huang, Gamini Dissanayake
Our SLAM system can: (1) Incrementally build a live model by progressively fusing new observations with vivid accurate texture.
Dynamic Reconstruction Simultaneous Localization and Mapping
no code implementations • 7 Apr 2020 • Qinghai Zheng, Jihua Zhu, Zhongyu Li, Shanmin Pang, Jun Wang, Lei Chen
The complementary graph regularizer investigates the specific information of multiple views.
no code implementations • ICML 2020 • Huajie Shao, Shuochao Yao, Dachun Sun, Aston Zhang, Shengzhong Liu, Dongxin Liu, Jun Wang, Tarek Abdelzaher
Variational Autoencoders (VAE) and their variants have been widely used in a variety of applications, such as dialog generation, image generation and disentangled representation learning.
1 code implementation • 23 Apr 2020 • Jingwei Song, Shaobo Xia, Jun Wang, Mitesh Patel, Dong Chen
Sliding-window based low-rank matrix approximation (LRMA) is a technique widely used in hyperspectral images (HSIs) denoising or completion.
no code implementations • 24 Apr 2020 • Dening Lu, Xuequan Lu, Yangxing Sun, Jun Wang
In this paper, we propose a novel feature-preserving normal estimation method for point cloud filtering with preserving geometric features.
no code implementations • 29 Apr 2020 • Minghao Han, Lixian Zhang, Jun Wang, Wei Pan
Reinforcement Learning (RL) and its integration with deep learning have achieved impressive performance in various robotic control tasks, ranging from motion planning and navigation to end-to-end visual manipulation.
no code implementations • 15 May 2020 • Zhe Wang, Jun Wang, Yezhou Yang
Pedestrian detection has been heavily studied in the last decade due to its wide application.
no code implementations • 19 May 2020 • Yuang Liu, Wei zhang, Jun Wang
Knowledge Distillation (KD) is an effective framework for compressing deep learning models, realized by a student-teacher paradigm requiring small student networks to mimic the soft target generated by well-trained teachers.
no code implementations • 26 May 2020 • Cong Wang, Yanru Xiao, Xing Gao, Li Li, Jun Wang
We show the feasibility of training with mobile CPUs, where training 100 epochs takes less than 10 mins and can be boosted 3-5 times with feature transfer.
no code implementations • 18 Jun 2020 • Shuai Zhang, Xiaoyan Xin, Yang Wang, Yachong Guo, Qiuqiao Hao, Xianfeng Yang, Jun Wang, Jian Zhang, Bing Zhang, Wei Wang
The model provides automated recognition of given scans and generation of reports.
no code implementations • 29 Jun 2020 • Kai Zhang, Yaokang Zhu, Jun Wang, Jie Zhang, Hongyuan Zha
Graph neural networks are promising architecture for learning and inference with graph-structured data.
no code implementations • 7 Jul 2020 • Xinyuan Liu, Jihua Zhu, Qinghai Zheng, Zhongyu Li, Ruixin Liu, Jun Wang
More specifically, this novel loss function not only considers the mapping errors generated from the projection of the input space into the output one but also accounts for the reconstruction errors generated from the projection of the output space back to the input one.
no code implementations • ECCV 2020 • Jun Wang, Shiyi Lan, Mingfei Gao, Larry S. Davis
Results show that our framework achieves the state-of-the-art performance with 31 FPS and improves our baseline significantly by 9. 0% mAP on the nuScenes test set.
Ranked #328 on 3D Object Detection on nuScenes
no code implementations • 20 Jul 2020 • Dan Zeng, Hailin Shi, Hang Du, Jun Wang, Zhen Lei, Tao Mei
However, the correlation between hard positive and hard negative is overlooked, and so is the relation between the margins in positive and negative logits.
no code implementations • 10 Aug 2020 • Longkang Li, Hui-Ling Zhen, Mingxuan Yuan, Jiawen Lu, XialiangTong, Jia Zeng, Jun Wang, Dirk Schnieders
In this paper, we propose a Bilevel Deep reinforcement learning Scheduler, \textit{BDS}, in which the higher level is responsible for exploring an initial global sequence, whereas the lower level is aiming at exploitation for partial sequence refinements, and the two levels are connected by a sliding-window sampling mechanism.
no code implementations • ICLR 2020 • Kai Zhang, Yaokang Zhu, Jun Wang, Jie Zhang
Yet, how to fully exploit rich structural information in the attention mechanism remains a challenge.
no code implementations • 3 Sep 2020 • Zhaoqing Peng, Junqi Jin, Lan Luo, Yaodong Yang, Rui Luo, Jun Wang, Wei-Nan Zhang, Haiyang Xu, Miao Xu, Chuan Yu, Tiejian Luo, Han Li, Jian Xu, Kun Gai
To drive purchase in online advertising, it is of the advertiser's great interest to optimize the sequential advertising strategy whose performance and interpretability are both important.
no code implementations • 3 Sep 2020 • Bin Huang, Yuanyang Du, Shuai Zhang, Wenfei Li, Jun Wang, Jian Zhang
RNAs play crucial and versatile roles in biological processes.
no code implementations • 15 Sep 2020 • Yuanpeng Liu, Jun Zhou, Yuqi Zhang, Chao Ding, Jun Wang
To address the problem, a novel 3DPVNet is presented in this work, which utilizes 3D local patches to vote for the object 6D poses.
no code implementations • 1 Jan 2021 • Minghao Han, Zhipeng Zhou, Lixian Zhang, Jun Wang, Wei Pan
Reinforcement learning is promising to control dynamical systems for which the traditional control methods are hardly applicable.
no code implementations • 1 Jan 2021 • Jiarui Jin, Cong Chen, Ming Zhou, Weinan Zhang, Rasool Fakoor, David Wipf, Yong Yu, Jun Wang, Alex Smola
Goal-oriented reinforcement learning algorithms are often good at exploration, not exploitation, while episodic algorithms excel at exploitation, not exploration.
no code implementations • 1 Jan 2021 • Yizheng Hu, Kun Shao, Dong Li, Jianye Hao, Wulong Liu, Yaodong Yang, Jun Wang, Zhanxing Zhu
Therefore, to achieve robust CMARL, we introduce novel strategies to encourage agents to learn correlated equilibrium while maximally preserving the convenience of the decentralized execution.
no code implementations • 1 Jan 2021 • Peng Zhang, Jing Zhang, Xindian Ma, Siwei Rao, Guangjian Tian, Jun Wang
As a novel model that bridges machine learning and quantum theory, tensor network (TN) has recently gained increasing attention and successful applications for processing natural images.
no code implementations • 1 Jan 2021 • Yean Hoon Ong, Jun Wang
Exploration is a long-standing challenge in sequential decision problem in machine learning.
no code implementations • 1 Jan 2021 • Yali Du, Yifan Zhao, Meng Fang, Jun Wang, Gangyan Xu, Haifeng Zhang
Dealing with multi-agent control in networked systems is one of the biggest challenges in Reinforcement Learning (RL) and limited success has been presented compared to recent deep reinforcement learning in single-agent domain.
no code implementations • 1 Jan 2021 • Peng Zhang, Furui Liu, Zhitang Chen, Jianye Hao, Jun Wang
Reinforcement Learning (RL) has shown great potential to deal with sequential decision-making problems.
no code implementations • 2 Oct 2020 • Shaowei Wei, Jun Wang, Guoxian Yu, Carlotta Domeniconi, Xiangliang Zhang
Multi-view clustering aims at exploiting information from multiple heterogeneous views to promote clustering.
no code implementations • 6 Oct 2020 • Yuanlin Yang, Guoxian Yu, Jun Wang, Carlotta Domeniconi, Xiangliang Zhang
Multi-typed objects Multi-view Multi-instance Multi-label Learning (M4L) deals with interconnected multi-typed objects (or bags) that are made of diverse instances, represented with heterogeneous feature views and annotated with a set of non-exclusive but semantically related labels.
no code implementations • 9 Oct 2020 • Jin Cao, Jun Wang, Wael Hamza, Kelly Vanee, Shang-Wen Li
The light encoder architecture separates the shared pre-trained networks from the mappings of generally encoded knowledge to specific domains of SLU, allowing for the domain adaptation to be performed solely at the light encoder and thus increasing efficiency.
no code implementations • 10 Oct 2020 • Jun Wang, Qianying Liu, Haotian Xie, Zhaogang Yang, Hefeng Zhou
In this paper, the Convolutional Neutral Network (CNN) has been adapted to predict and classify lymph node metastasis in breast cancer.
no code implementations • 18 Feb 2020 • Jun Wang
The author is mainly interest in the Gr\"{o}bner-Shirshov bases of finite Coxeter groups.
Group Theory Rings and Algebras 16S15, 20F55
no code implementations • 30 Aug 2020 • Aleks Jevnikar, Jun Wang, Wen Yang
In the present paper we derive Liouville type results and existence of periodic solutions for $\chi^{(2)}$ type systems with non-homogeneous nonlinearities.
Analysis of PDEs 35K9, 35J61, 35B45
no code implementations • 2 Nov 2020 • Chang Xu, Jun Wang, Yuqing Tang, Francisco Guzman, Benjamin I. P. Rubinstein, Trevor Cohn
In this paper, we show that targeted attacks on black-box NMT systems are feasible, based on poisoning a small fraction of their parallel training data.
no code implementations • 1 Nov 2020 • Xinyi Dai, Jiawei Hou, Qing Liu, Yunjia Xi, Ruiming Tang, Weinan Zhang, Xiuqiang He, Jun Wang, Yong Yu
To this end, we propose a novel ranking framework called U-rank that directly optimizes the expected utility of the ranking list.
no code implementations • NeurIPS 2020 • Rui Luo, Qiang Zhang, Yaodong Yang, Jun Wang
In this paper, we present a new practical method for Bayesian learning that can rapidly draw representative samples from complex posterior distributions with multiple isolated modes in the presence of mini-batch noise.
no code implementations • 13 Nov 2020 • Minghao Han, Yuan Tian, Lixian Zhang, Jun Wang, Wei Pan
In comparison with the existing RL algorithms, the proposed method can achieve superior performance in terms of maintaining safety.
no code implementations • 29 Nov 2020 • Haotian Xie, Yong Zhang, Jun Wang, Jingjing Zhang, Yifan Ma, Zhaogang Yang
The Gleason grading system using histological images is the most powerful diagnostic and prognostic predictor of prostate cancer.
no code implementations • 28 Nov 2019 • Guoyong Zhang, Xiao Fu, Jun Wang, Xi-Le Zhao, Mingyi Hong
Spectrum cartography aims at estimating power propagation patterns over a geographical region across multiple frequency bands (i. e., a radio map)---from limited samples taken sparsely over the region.
no code implementations • 18 Jun 2020 • Meng Ding, Xiao Fu, Ting-Zhu Huang, Jun Wang, Xi-Le Zhao
This work employs an idea that models spectral images as tensors following the block-term decomposition model with multilinear rank-$(L_r, L_r, 1)$ terms (i. e., the LL1 model) and formulates the HSR problem as a coupled LL1 tensor decomposition problem.
no code implementations • 19 Aug 2020 • Wenpo Yao, Jun Wang, Matjaz Perc, Wenli Yao, Jiafei Dai, Daqing Guo, Dezhong Yao
Time irreversibility should be measured based on the permutations of symmetric vectors rather than symmetric permutations, whereas symmetric permutations can instead be employed to determine the quantitative amplitude irreversibility -- a novel parameter proposed in this paper for nonequilibrium calculated by means of the probabilistic difference in amplitude fluctuations.
no code implementations • 9 Dec 2020 • Jun Wang, Shaoguo Wen, Kaixing Chen, Jianghua Yu, Xin Zhou, Peng Gao, Changsheng Li, Guotong Xie
Active learning generally involves querying the most representative samples for human labeling, which has been widely studied in many fields such as image classification and object detection.
no code implementations • 11 Dec 2020 • Fitash Ul Haq, Donghwan Shin, Lionel C. Briand, Thomas Stifter, Jun Wang
In this paper, we present an approach to automatically generate test data for KP-DNNs using many-objective search.
no code implementations • 21 Dec 2020 • Pengyong Li, Jun Wang, Yixuan Qiao, Hao Chen, Yihuan Yu, Xiaojun Yao, Peng Gao, Guotong Xie, Sen Song
Here, we proposed a novel Molecular Pre-training Graph-based deep learning framework, named MPG, that leans molecular representations from large-scale unlabeled molecules.
no code implementations • 22 Dec 2020 • Yao Zhang, Xu Zhang, Jun Wang, Hongru Liang, Wenqiang Lei, Zhe Sun, Adam Jatowt, Zhenglu Yang
The current methods for the link prediction taskhavetwonaturalproblems:1)the relation distributions in KGs are usually unbalanced, and 2) there are many unseen relations that occur in practical situations.
no code implementations • 28 Dec 2020 • Minne Li, Mengyue Yang, Furui Liu, Xu Chen, Zhitang Chen, Jun Wang
The capability of imagining internally with a mental model of the world is vitally important for human cognition.
no code implementations • 15 Jan 2021 • Ronglin Gong, Jun Wang, Jun Shi
In this work, a Task-driven Self-supervised Bi-channel Networks (TSBN) framework is proposed to improve the performance of classification model the mammography-based CAD.
no code implementations • 15 Jan 2021 • Vincent Moens, Hang Ren, Alexandre Maraval, Rasul Tutunov, Jun Wang, Haitham Ammar
In this paper, we propose CI-VI an efficient and scalable solver for semi-implicit variational inference (SIVI).
no code implementations • 17 Jan 2021 • Jinye Peng, Jiaxin Wang, Jun Wang, Erlei Zhang, Qunxi Zhang, Yongqin Zhang, Xianlin Peng, Kai Yu
For the fine extraction stage, we design a new multiscale U-Net (MSU-Net) to effectively remove disease noise and refine the sketch.
no code implementations • 15 Feb 2021 • Yaodong Yang, Jun Luo, Ying Wen, Oliver Slumbers, Daniel Graves, Haitham Bou Ammar, Jun Wang, Matthew E. Taylor
Multiagent reinforcement learning (MARL) has achieved a remarkable amount of success in solving various types of video games.
no code implementations • 23 Feb 2021 • Huajie Shao, Jun Wang, Haohong Lin, Xuezhou Zhang, Aston Zhang, Heng Ji, Tarek Abdelzaher
The algorithm is injected into a Conditional Variational Autoencoder (CVAE), allowing \textit{Apex} to control both (i) the order of keywords in the generated sentences (conditioned on the input keywords and their order), and (ii) the trade-off between diversity and accuracy.
no code implementations • 1 Mar 2021 • Jun Wang, Max W. Y. Lam, Dan Su, Dong Yu
To extract robust deep representations from long sequential modeling of speech data, we propose a self-supervised learning approach, namely Contrastive Separative Coding (CSC).
no code implementations • 2 Mar 2021 • Jun Wang, Max W. Y. Lam, Dan Su, Dong Yu
We study the cocktail party problem and propose a novel attention network called Tune-In, abbreviated for training under negative environments with interference.
no code implementations • 7 Mar 2021 • Yunxiang Li, Yifan Zhang, Yaqi Wang, Shuai Wang, Ruizi Peng, Kai Tang, Qianni Zhang, Jun Wang, Qun Jin, Lingling Sun
As the most economical and routine auxiliary examination in the diagnosis of root canal treatment, oral X-ray has been widely used by stomatologists.
no code implementations • 16 Mar 2021 • David Mguni, Taher Jafferjee, Jianhong Wang, Nicolas Perez-Nieves, Tianpei Yang, Matthew Taylor, Wenbin Song, Feifei Tong, Hui Chen, Jiangcheng Zhu, Jun Wang, Yaodong Yang
Reward shaping (RS) is a powerful method in reinforcement learning (RL) for overcoming the problem of sparse or uninformative rewards.
no code implementations • 22 Mar 2021 • Jian Wang, Chen Xu, Rong Li, Yiqun Ge, Jun Wang
We not only verify the performance gain achieved, but also provide implementation-friend designs, i. e., a scalable neural network design for the agent and a virtual environment training framework.
no code implementations • 23 Mar 2021 • Hongru Liang, Haozheng Wang, Qian Li, Jun Wang, Guandong Xu, Jiawei Chen, Jin-Mao Wei, Zhenglu Yang
Learning and analyzing rap lyrics is a significant basis for many web applications, such as music recommendation, automatic music categorization, and music information retrieval, due to the abundant source of digital music in the World Wide Web.
no code implementations • 24 Mar 2021 • Yishan He, Fei Gao, Jun Wang, Amir Hussain, Erfu Yang, Huiyu Zhou
In this paper, in order to solve the boundary discontinuity problem in OBB regression, we propose to detect SAR ships by learning polar encodings.
no code implementations • CVPR 2021 • Yuang Liu, Wei zhang, Jun Wang
To address the above issues, we propose a zero-shot adversarial quantization (ZAQ) framework, facilitating effective discrepancy estimation and knowledge transfer from a full-precision model to its quantized model.
Ranked #2 on Data Free Quantization on CIFAR-100 (CIFAR-100 W5A5 Top-1 Accuracy metric)
no code implementations • CVPR 2021 • Yuang Liu, Wei zhang, Jun Wang
To cope with this issue, we propose a source-free domain adaptation framework for semantic segmentation, namely SFDA, in which only a well-trained source model and an unlabeled target domain dataset are available for adaptation.
no code implementations • 1 Apr 2021 • Jun Wang
We explore the effectiveness of some representative detectors utilized as the second-phase detector in two-phase WSOD and propose a two-phase WSOD architecture.
no code implementations • 5 Apr 2021 • Jiyuan Hu, Jun Wang, Guangyu Zhong, Jian Cao, Ren Mao, Fan Liang
The reference frame memory accesses in inter prediction result in high DRAM bandwidth requirement and power consumption.
no code implementations • 3 Dec 2020 • Yingzhuo Sun, Cheng-Xiang Wang, Jie Huang, Jun Wang
The evolution of clusters on the linear array and planar array is also considered in the proposed model.
no code implementations • 15 Aug 2020 • Lei LI, Qihang Peng, Jun Wang
The first is based on multi-view convolutional neural network by treating signals from different receive antennas as different views of a 3D object and designing the location and operation of view-pooling layer that are suitable for feature fusion of multi-antenna signals.
no code implementations • 15 Apr 2021 • Michael Neumann, Oliver Roesler, Jackson Liscombe, Hardik Kothare, David Suendermann-Oeft, David Pautler, Indu Navar, Aria Anvar, Jochen Kumm, Raquel Norel, Ernest Fraenkel, Alexander V. Sherman, James D. Berry, Gary L. Pattee, Jun Wang, Jordan R. Green, Vikram Ramanarayanan
Our results provide encouraging evidence of the utility of automatically extracted audiovisual analytics for scalable remote patient assessment and monitoring in ALS.
no code implementations • 20 Apr 2021 • Jun Wang, Cheng-Xiang Wang, Jie Huang, Haiming Wang, Xiqi Gao
The proposed THz channel model is very general having the capability to capture different channel characteristics in multiple THz application scenarios such as indoor scenarios, device-to-device (D2D) communications, ultra-massive multiple-input multiple-output (MIMO) communications, and long traveling paths of users.
no code implementations • 25 Apr 2021 • Yingzhuo Sun, Cheng-Xiang Wang, Jie Huang, Jun Wang
In this paper, a three-dimensional (3D) geometry based stochastic model (GBSM) for a massive multiple-input multiple-output (MIMO) communication system employing practical discrete intelligent reflecting surface (IRS) is proposed.
1 code implementation • 2 May 2021 • Yunxiang Li, Guodong Zeng, Yifan Zhang, Jun Wang, Qianni Zhang, Qun Jin, Lingling Sun, Qisi Lian, Neng Xia, Ruizi Peng, Kai Tang, Yaqi Wang, Shuai Wang
Accurate evaluation of the treatment result on X-ray images is a significant and challenging step in root canal therapy since the incorrect interpretation of the therapy results will hamper timely follow-up which is crucial to the patients' treatment outcome.
no code implementations • 7 May 2021 • Yiming Bao, Jun Wang, Tong Li, Linyan Wang, Jianwei Xu, Juan Ye, Dahong Qian
Specifically, the encoder of a DL model that is pre-trained on the source domain is used to initialize the encoder of a reconstruction model.
no code implementations • 19 May 2021 • Yu Wang, Hejia Luo, Ying Chen, Jun Wang, Rong Li, Bin Wang
Next generation beyond 5G networks are expected to provide both Terabits per second data rate communication services and centimeter-level accuracy localization services in an efficient, seamless and cost-effective manner.
no code implementations • 27 May 2021 • Xijun Li, Weilin Luo, Mingxuan Yuan, Jun Wang, Jiawen Lu, Jie Wang, Jinhu Lu, Jia Zeng
Our method is entirely data driven and thus adaptive, i. e., the relational representation of adjacent vehicles can be learned and corrected by ST-DDGN from data periodically.
no code implementations • 28 May 2021 • Zeren Huang, Kerong Wang, Furui Liu, Hui-Ling Zhen, Weinan Zhang, Mingxuan Yuan, Jianye Hao, Yong Yu, Jun Wang
In the online A/B testing of the product planning problems with more than $10^7$ variables and constraints daily, Cut Ranking has achieved the average speedup ratio of 12. 42% over the production solver without any accuracy loss of solution.
no code implementations • 8 Jun 2021 • Max W. Y. Lam, Jun Wang, Chao Weng, Dan Su, Dong Yu
End-to-end speech recognition generally uses hand-engineered acoustic features as input and excludes the feature extraction module from its joint optimization.
no code implementations • NAACL 2021 • Xiyan Fu, Jun Wang, Zhenglu Yang
Multimodal summarization becomes increasingly significant as it is the basis for question answering, Web search, and many other downstream tasks.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Zhen Wang, Siwei Rao, Jie Zhang, Zhen Qin, Guangjian Tian, Jun Wang
However, question generation is actually a one-to-many problem, as it is possible to raise questions with different focuses on contexts and various means of expression.
no code implementations • 24 Jun 2021 • Yixuan Qiao, Hao Chen, Jun Wang, Yihao Chen, Xianbin Ye, Ziliang Li, Xianbiao Qi, Peng Gao, Guotong Xie
TextVQA requires models to read and reason about text in images to answer questions about them.
no code implementations • 29 Jun 2021 • Zhiyang Lu, Zheng Li, Jun Wang, Jun Shi, Dinggang Shen
To this end, we propose a novel Two-stage Self-supervised Cycle-consistency Network (TSCNet) for MR slice interpolation, in which a two-stage self-supervised learning (SSL) strategy is developed for unsupervised DL network training.
no code implementations • 6 Jul 2021 • Vincent Moens, Aivar Sootla, Haitham Bou Ammar, Jun Wang
We present a method for conditional sampling for pre-trained normalizing flows when only part of an observation is available.
1 code implementation • 12 Jul 2021 • Jun Wang, Chang Xu, Francisco Guzman, Ahmed El-Kishky, Yuqing Tang, Benjamin I. P. Rubinstein, Trevor Cohn
Neural machine translation systems are known to be vulnerable to adversarial test inputs, however, as we show in this paper, these systems are also vulnerable to training attacks.
1 code implementation • 14 Jul 2021 • Jun Wang, Bei Yu
Accurately linking news articles to scientific research works is a critical component in a number of applications, such as measuring the social impact of a research work and detecting inaccuracies or distortions in science news.
1 code implementation • Findings (ACL) 2021 • Jun Wang, Chang Xu, Francisco Guzman, Ahmed El-Kishky, Benjamin I. P. Rubinstein, Trevor Cohn
Mistranslated numbers have the potential to cause serious effects, such as financial loss or medical misinformation.
no code implementations • 31 Jul 2021 • Jian Wang, Yourui Huangfu, Rong Li, Yiqun Ge, Jun Wang
The wireless network is undergoing a trend from "onnection of things" to "connection of intelligence".
no code implementations • 14 Aug 2021 • Jun Wang, Cheng-Xiang Wang, Jie Huang, Haiming Wang, Xiqi Gao, Xiaohu You, Yang Hao
Terahertz (THz) communication is now being considered as one of possible technologies for the sixth generation (6G) wireless communication systems.
no code implementations • 17 Aug 2021 • Zhijian Duan, Wenhan Huang, Dinghuai Zhang, Yali Du, Jun Wang, Yaodong Yang, Xiaotie Deng
In this paper, we investigate the learnability of the function approximator that approximates Nash equilibrium (NE) for games generated from a distribution.
no code implementations • 24 Aug 2021 • Xidong Feng, Chen Chen, Dong Li, Mengchen Zhao, Jianye Hao, Jun Wang
Meta learning, especially gradient based one, can be adopted to tackle this problem by learning initial parameters of the model and thus allowing fast adaptation to a specific task from limited data examples.
no code implementations • 26 Aug 2021 • Max W. Y. Lam, Jun Wang, Rongjie Huang, Dan Su, Dong Yu
In this paper, we propose novel bilateral denoising diffusion models (BDDMs), which take significantly fewer steps to generate high-quality samples.
no code implementations • 25 Aug 2021 • Jun Wang, Hefeng Zhou, Xiaohan Yu
There are two main problems hindering the performance of the two-phase WSOD approaches, i. e., insufficient learning problem and strict reliance between the FSD and the pseudo ground truth (PGT) generated by the WSOD model.
no code implementations • 2 Sep 2021 • Mengyue Yang, Quanyu Dai, Zhenhua Dong, Xu Chen, Xiuqiang He, Jun Wang
To alleviate this problem, in this paper, we propose to reformulate the recommendation task within the causal inference framework, which enables us to counterfactually simulate user ranking-based preferences to handle the data scarce problem.
no code implementations • 4 Sep 2021 • Xiaotie Deng, Ningyuan Li, David Mguni, Jun Wang, Yaodong Yang
Similar to the role of Markov decision processes in reinforcement learning, Stochastic Games (SGs) lay the foundation for the study of multi-agent reinforcement learning (MARL) and sequential agent interactions.
Multi-agent Reinforcement Learning reinforcement-learning +1
no code implementations • 5 Sep 2021 • Dell Zhang, Jun Wang
Driven by the need to capture users' evolving interests and optimize their long-term experiences, more and more recommender systems have started to model recommendation as a Markov decision process and employ reinforcement learning to address the problem.
no code implementations • 20 Sep 2021 • Yixin Wu, Rui Luo, Chen Zhang, Jun Wang, Yaodong Yang
In this paper, we characterize the noise of stochastic gradients and analyze the noise-induced dynamics during training deep neural networks by gradient-based optimizers.
no code implementations • ICCV 2021 • Qian Xie, Yu-Kun Lai, Jing Wu, Zhoutao Wang, Dening Lu, Mingqiang Wei, Jun Wang
Hough voting, as has been demonstrated in VoteNet, is effective for 3D object detection, where voting is a key step.
no code implementations • ICCV 2021 • Yajing Kong, Liu Liu, Jun Wang, DaCheng Tao
Therefore, in contrast to recent works using a fixed curriculum, we devise a new curriculum learning method, Adaptive Curriculum Learning (Adaptive CL), adapting the difficulty of examples to the current state of the model.
no code implementations • 7 Oct 2021 • Le Cong Dinh, David Henry Mguni, Long Tran-Thanh, Jun Wang, Yaodong Yang
In this setting, we first demonstrate that MDP-Expert, an existing algorithm that works well with oblivious adversaries can still apply and achieve a policy regret bound of $\mathcal{O}(\sqrt{T \log(L)}+\tau^2\sqrt{ T \log(|A|)})$ where $L$ is the size of adversary's pure strategy set and $|A|$ denotes the size of agent's action space.
no code implementations • 29 Sep 2021 • Rongjie Huang, Max W. Y. Lam, Jun Wang, Dan Su, Dong Yu, Zhou Zhao, Yi Ren
Learning generalizable speech representations for unseen samples in different domains has been a challenge with ever increasing importance to date.
no code implementations • 29 Sep 2021 • Linghui Meng, Muning Wen, Yaodong Yang, Chenyang Le, Xi yun Li, Haifeng Zhang, Ying Wen, Weinan Zhang, Jun Wang, Bo Xu
Offline reinforcement learning leverages static datasets to learn optimal policies with no necessity to access the environment.
Multi-agent Reinforcement Learning reinforcement-learning +2
no code implementations • 29 Sep 2021 • Jing Zhang, Peng Zhang, Yupeng He, Siwei Rao, Jun Wang, Guangjian Tian
In this framework, we derive the mathematical representation of the variable space, and then use a tensor network based on the idea of low-rank approximation to model the variable space.
no code implementations • 29 Sep 2021 • Guochang Lin, Fukai Chen, Pipi Hu, Xiang Chen, Junqing Chen, Jun Wang, Zuoqiang Shi
Green's function plays a significant role in both theoretical analysis and numerical computing of partial differential equations (PDEs).
no code implementations • 29 Sep 2021 • Yuchen Liu, Yali Du, Runji Lin, Hangrui Bi, Mingdong Wu, Jun Wang, Hao Dong
Model-based RL is an effective approach for reducing sample complexity.
Model-based Reinforcement Learning Reinforcement Learning (RL)
no code implementations • 29 Sep 2021 • Mengyue Yang, Furui Liu, Xu Chen, Zhitang Chen, Jianye Hao, Jun Wang
In many real-world scenarios, such as image classification and recommender systems, it is evidence that representation learning can improve model's performance over multiple downstream tasks.
no code implementations • 29 Sep 2021 • Jun Wang, Tianxiang Su
In our experiments, we compared the proposed method with multiple state-of-the-art continual learning methods and the mREMIND network outperformed others both in accuracy and in disk space usage.
no code implementations • 29 Sep 2021 • Jun Wang, Yiannis Kantaros
To mitigate this challenge, in this paper, we propose model-based robust adaptive training algorithm (MRTAdapt), a new training algorithm to enhance the robustness of DNN-based semantic segmentation methods against natural variations that leverages model-based robust training algorithms and generative adversarial networks.
no code implementations • NeurIPS 2021 • Minghuan Liu, Zhengbang Zhu, Yuzheng Zhuang, Weinan Zhang, Jian Shen, Jianye Hao, Yong Yu, Jun Wang
State-only imitation learning (SOIL) enables agents to learn from massive demonstrations without explicit action or reward information.
no code implementations • 29 Sep 2021 • Hangyu Mao, Jianye Hao, Dong Li, Jun Wang, Weixun Wang, Xiaotian Hao, Bin Wang, Kun Shao, Zhen Xiao, Wulong Liu
In contrast, we formulate an \emph{explicit} credit assignment problem where each agent gives its suggestion about how to weight individual Q-values to explicitly maximize the joint Q-value, besides guaranteeing the Bellman optimality of the joint Q-value.
no code implementations • 21 Oct 2021 • Yanbin He, Zhiyang Lu, Jun Wang, Jun Shi
Convolutional neural networks (CNNs) and their variants have been successfully applied to the electroencephalogram (EEG) based motor imagery (MI) decoding task.
no code implementations • 22 Oct 2021 • Ricky Sanjaya, Jun Wang, Yaodong Yang
In this paper, we quantify the non-transitivity in Chess through real-world data from human players.
no code implementations • 26 Oct 2021 • Pengyong Li, Jun Wang, Ziliang Li, Yixuan Qiao, Xianggen Liu, Fei Ma, Peng Gao, Seng Song, Guotong Xie
Self-supervised learning has gradually emerged as a powerful technique for graph representation learning.
no code implementations • 27 Oct 2021 • David Mguni, Usman Islam, Yaqi Sun, Xiuling Zhang, Joel Jennings, Aivar Sootla, Changmin Yu, Ziyan Wang, Jun Wang, Yaodong Yang
In this paper, we introduce a new generation of RL solvers that learn to minimise safety violations while maximising the task reward to the extent that can be tolerated by the safe policy.
no code implementations • 28 Oct 2021 • Yunxiang Li, Jingxiong Li, Ruilong Dan, Shuai Wang, Kai Jin, Guodong Zeng, Jun Wang, Xiangji Pan, Qianni Zhang, Huiyu Zhou, Qun Jin, Li Wang, Yaqi Wang
To mitigate this problem, a novel unsupervised domain adaptation (UDA) method named dispensed Transformer network (DTNet) is introduced in this paper.
no code implementations • 28 Oct 2021 • Chenguang Wang, Yaodong Yang, Oliver Slumbers, Congying Han, Tiande Guo, Haifeng Zhang, Jun Wang
In this paper, we introduce a two-player zero-sum framework between a trainable \emph{Solver} and a \emph{Data Generator} to improve the generalization ability of deep learning-based solvers for Traveling Salesman Problem (TSP).