no code implementations • Findings (EMNLP) 2021 • Chang Xu, Jun Wang, Francisco Guzmán, Benjamin Rubinstein, Trevor Cohn
NLP models are vulnerable to data poisoning attacks.
no code implementations • ACL 2022 • Jun Wang, Benjamin Rubinstein, Trevor Cohn
In this paper we describe a new source of bias prevalent in NMT systems, relating to translations of sentences containing person names.
no code implementations • ACL 2022 • Ling.Yu Zhu, Zhengkun Zhang, Jun Wang, Hongbin Wang, Haiying Wu, Zhenglu Yang
Empathetic dialogue assembles emotion understanding, feeling projection, and appropriate response generation.
no code implementations • CCL 2022 • Zekun Deng, Hao Yang, Jun Wang
"《史记》和《汉书》具有经久不衰的研究价值。尽管两书异同的研究已经较为丰富, 但研究的全面性、完备性、科学性、客观性均仍显不足。在数字人文的视角下, 本文利用计算语言学方法, 通过对字、词、命名实体、段落等的多粒度、多角度分析, 开展对于《史》《汉》的比较研究。首先, 本文对于《史》《汉》中的字、词、命名实体的分布和特点进行对比, 以遍历穷举的考察方式提炼出两书在主要内容上的相同点与不同点, 揭示了汉武帝之前和汉武帝到西汉灭亡两段历史时期在政治、文化、思想上的重要变革与承袭。其次, 本文使用一种融入命名实体作为外部特征的文本相似度算法对于《史记》《汉书》的异文进行自动发现, 成功识别出过去研究者通过人工手段没有发现的袭用段落, 使得我们对于《史》《汉》的承袭关系形成更加完整和立体的认识。再次, 本文通过计算异文段落之间的最长公共子序列来自动得出两段异文之间存在的差异, 从宏观统计上证明了《汉书》文字风格《史记》的差别, 并从微观上进一步对二者语言特点进行了阐释, 为理解《史》《汉》异文特点提供了新的角度和启发。本研究站在数字人文的视域下, 利用先进的计算方法对于传世千年的中国古代经典进行了再审视、再发现, 其方法对于今人研究古籍有一定的借鉴价值。”
no code implementations • COLING 2022 • Alex X. Zhang, Xun Liang, Bo Wu, Xiangping Zheng, Sensen Zhang, Yuhui Guo, Jun Wang, Xinyao Liu
The human recognition system has presented the remarkable ability to effortlessly learn novel knowledge from only a few trigger events based on prior knowledge, which is called insight learning.
1 code implementation • EMNLP 2021 • Yingya Li, Jun Wang, Bei Yu
We also conducted a case study that applied this prediction model to retrieve specific health advice on COVID-19 treatments from LitCovid, a large COVID research literature portal, demonstrating the usefulness of retrieving health advice sentences as an advanced research literature navigation function for health researchers and the general public.
no code implementations • SemEval (NAACL) 2022 • Qizhi Lin, Changyu Hou, Xiaopeng Wang, Jun Wang, Yixuan Qiao, Peng Jiang, Xiandi Jiang, Benqi Wang, Qifeng Xiao
From pretrained contextual embedding to document-level embedding, the selection and construction of embedding have drawn more and more attention in the NER domain in recent research.
no code implementations • 3 Jun 2023 • Xuemei Tang, Jun Wang, Qi Su
Recently, it is quite common to integrate Chinese sequence labeling results to enhance syntactic and semantic parsing.
no code implementations • 28 May 2023 • Yudi Zhang, Yali Du, Biwei Huang, Ziyan Wang, Jun Wang, Meng Fang, Mykola Pechenizkiy
While the majority of current approaches construct the reward redistribution in an uninterpretable manner, we propose to explicitly model the contributions of state and action from a causal perspective, resulting in an interpretable return decomposition.
no code implementations • 26 May 2023 • Yingjie Feng, Jun Wang, Xianfeng GU, Xiaoyin Xu, Min Zhang
In diagnosing challenging conditions such as Alzheimer's disease (AD), imaging is an important reference.
1 code implementation • 25 May 2023 • Xuanli He, Jun Wang, Benjamin Rubinstein, Trevor Cohn
Backdoor attacks are an insidious security threat against machine learning models.
no code implementations • 25 May 2023 • Wuwei Lan, Zhiguo Wang, Anuj Chauhan, Henghui Zhu, Alexander Li, Jiang Guo, Sheng Zhang, Chung-Wei Hang, Joseph Lilien, Yiqun Hu, Lin Pan, Mingwen Dong, Jun Wang, Jiarong Jiang, Stephen Ash, Vittorio Castelli, Patrick Ng, Bing Xiang
A practical text-to-SQL system should generalize well on a wide variety of natural language questions, unseen database schemas, and novel SQL query structures.
no code implementations • 25 May 2023 • Saisai Ding, Juncheng Li, Jun Wang, Shihui Ying, Jun Shi
The key idea of MEGT is to adopt two independent Efficient Graph-based Transformer (EGT) branches to process the low-resolution and high-resolution patch embeddings (i. e., tokens in a Transformer) of WSIs, respectively, and then fuse these tokens via a multi-scale feature fusion module (MFFM).
no code implementations • 24 May 2023 • Zheng Hu, Shi-Min Cai, Jun Wang, Tao Zhou
Thus, the representation of users' dislikes should be integrated into the user modelling when we construct a collaborative recommendation model.
1 code implementation • 23 May 2023 • Wei Liu, Jun Wang, Haozhao Wang, Ruixuan Li, Yang Qiu, Yuankai Zhang, Jie Han, Yixiong Zou
However, such a cooperative game may incur the degeneration problem where the predictor overfits to the uninformative pieces generated by a not yet well-trained generator and in turn, leads the generator to converge to a sub-optimal model that tends to select senseless pieces.
no code implementations • 19 May 2023 • Xuanli He, Qiongkai Xu, Jun Wang, Benjamin Rubinstein, Trevor Cohn
Modern NLP models are often trained over large untrusted datasets, raising the potential for a malicious adversary to compromise model behaviour.
no code implementations • 16 May 2023 • Desong Du, Shaohang Han, Naiming Qi, Haitham Bou Ammar, Jun Wang, Wei Pan
Reinforcement learning (RL) exhibits impressive performance when managing complicated control tasks for robots.
1 code implementation • 16 May 2023 • Yan Song, He Jiang, Zheng Tian, Haifeng Zhang, Yingping Zhang, Jiangcheng Zhu, Zonghong Dai, Weinan Zhang, Jun Wang
Few multi-agent reinforcement learning (MARL) research on Google Research Football (GRF) focus on the 11v11 multi-agent full-game scenario and to the best of our knowledge, no open benchmark on this scenario has been released to the public.
no code implementations • 12 May 2023 • Jian Zhao, Jianan Li, Lei Jin, Jiaming Chu, Zhihao Zhang, Jun Wang, Jiangqiang Xia, Kai Wang, Yang Liu, Sadaf Gulshad, Jiaojiao Zhao, Tianyang Xu, XueFeng Zhu, Shihan Liu, Zheng Zhu, Guibo Zhu, Zechao Li, Zheng Wang, Baigui Sun, Yandong Guo, Shin ichi Satoh, Junliang Xing, Jane Shen Shengmei
Second, we set up two tracks for the first time, i. e., Anti-UAV Tracking and Anti-UAV Detection & Tracking.
1 code implementation • 8 May 2023 • Wei Liu, Haozhao Wang, Jun Wang, Ruixuan Li, Xinyang Li, Yuankai Zhang, Yang Qiu
Rationalization is to employ a generator and a predictor to construct a self-explaining NLP model in which the generator selects a subset of human-intelligible pieces of the input text to the following predictor.
no code implementations • 5 May 2023 • Yiyi Zhang, Zhiwen Ying, Ying Zheng, Cuiling Wu, Nannan Li, Jun Wang, Xianzhong Feng, Xiaogang Xu
Plant leaf identification is crucial for biodiversity protection and conservation and has gradually attracted the attention of academia in recent years.
no code implementations • 26 Apr 2023 • Xiaorui Wang, Jun Wang, Xin Tang, Peng Gao, Rui Fang, Guotong Xie
Filter pruning is widely adopted to compress and accelerate the Convolutional Neural Networks (CNNs), but most previous works ignore the relationship between filters and channels in different layers.
no code implementations • 26 Apr 2023 • Meixuan Qiao, Jun Wang, Junfu Xiang, Qiyu Hou, Ruixuan Li
Accurately extracting structured data from structure diagrams in financial announcements is of great practical importance for building financial knowledge graphs and further improving the efficiency of various financial applications.
no code implementations • 20 Apr 2023 • Yuyuan Li, Chaochao Chen, Xiaolin Zheng, Yizhao Zhang, Biao Gong, Jun Wang
In this paper, we first identify two main disadvantages of directly applying existing unlearning methods in the context of recommendation, i. e., (i) unsatisfactory efficiency for large-scale recommendation models and (ii) destruction of collaboration across users and items.
no code implementations • 11 Apr 2023 • Jun Wang, Omran Alamayreh, Benedetta Tondi, Mauro Barni
Classification of AI-manipulated content is receiving great attention, for distinguishing different types of manipulations.
1 code implementation • 28 Mar 2023 • Dan You, Pengcheng Xia, Qiuzhu Chen, Minghui Wu, Suncheng Xiang, Jun Wang
Automated chromosome instance segmentation from metaphase cell microscopic images is critical for the diagnosis of chromosomal disorders (i. e., karyotype analysis).
no code implementations • CVPR 2023 • Bo He, Jun Wang, JieLin Qiu, Trung Bui, Abhinav Shrivastava, Zhaowen Wang
The goal of multimodal summarization is to extract the most important information from different modalities to form summaries.
Ranked #3 on
Supervised Video Summarization
on SumMe
Extractive Text Summarization
Supervised Video Summarization
no code implementations • 12 Mar 2023 • Jun Wang, Klaus Mueller
Furthermore, since an effect can be a cause of other effects, we allow users to aggregate different temporal cause-effect relations found with our method into a visual flow diagram to enable the discovery of temporal causal networks.
1 code implementation • CVPR 2023 • Suhang Ye, Yingyi Zhang, Jie Hu, Liujuan Cao, Shengchuan Zhang, Lei Shen, Jun Wang, Shouhong Ding, Rongrong Ji
Specifically, DistilPose maximizes the transfer of knowledge from the teacher model (heatmap-based) to the student model (regression-based) through Token-distilling Encoder (TDE) and Simulated Heatmaps.
no code implementations • 13 Feb 2023 • Xihuai Wang, Zheng Tian, Ziyu Wan, Ying Wen, Jun Wang, Weinan Zhang
In this paper, we propose the \textbf{A}gent-by-\textbf{a}gent \textbf{P}olicy \textbf{O}ptimization (A2PO) algorithm to improve the sample efficiency and retain the guarantees of monotonic improvement for each agent during training.
no code implementations • 10 Feb 2023 • Jun Wang, Suyi Li
In this paper, we experimentally examine the cognitive capability of a simple, paper-based Miura-ori -- using the physical reservoir computing framework -- to achieve different information perception tasks.
no code implementations • 4 Feb 2023 • Jun Wang, Yue Song, David John Hill, Yunhe Hou, Feilong Fan
To figure out the stability issues brought by renewable energy sources (RES) with non-Gaussian uncertainties in isolated microgrids, this paper proposes a chance constrained stability constrained optimal power flow (CC-SC-OPF) model.
no code implementations • 21 Jan 2023 • Shuaichen Chang, Jun Wang, Mingwen Dong, Lin Pan, Henghui Zhu, Alexander Hanbo Li, Wuwei Lan, Sheng Zhang, Jiarong Jiang, Joseph Lilien, Steve Ash, William Yang Wang, Zhiguo Wang, Vittorio Castelli, Patrick Ng, Bing Xiang
Neural text-to-SQL models have achieved remarkable performance in translating natural language questions into SQL queries.
no code implementations • 19 Jan 2023 • Chengjie Zhao, Jun Wang, Wei Huang, Xiaonan Chen, Tianfu Qi
Under MGIN channel, classical communication signal schemes and corresponding detection methods usually can not achieve desirable performance as they are optimized with respect to WGN.
1 code implementation • 16 Jan 2023 • Xingzhou Lou, Jiaxian Guo, Junge Zhang, Jun Wang, Kaiqi Huang, Yali Du
We conduct experiments on the Overcooked environment, and evaluate the zero-shot human-AI coordination performance of our method with both behavior-cloned human proxies and real humans.
1 code implementation • 24 Dec 2022 • Ying Wen, Ziyu Wan, Ming Zhou, Shufang Hou, Zhe Cao, Chenyang Le, Jingxiao Chen, Zheng Tian, Weinan Zhang, Jun Wang
The pervasive uncertainty and dynamic nature of real-world environments present significant challenges for the widespread implementation of machine-driven Intelligent Decision-Making (IDM) systems.
no code implementations • 17 Dec 2022 • Yiyun Zhao, Jiarong Jiang, Yiqun Hu, Wuwei Lan, Henry Zhu, Anuj Chauhan, Alexander Li, Lin Pan, Jun Wang, Chung-Wei Hang, Sheng Zhang, Marvin Dong, Joe Lilien, Patrick Ng, Zhiguo Wang, Vittorio Castelli, Bing Xiang
In this paper, we first examined the existing synthesized datasets and discovered that state-of-the-art text-to-SQL algorithms did not further improve on popular benchmarks when trained with augmented synthetic data.
no code implementations • 15 Dec 2022 • Hang Lai, Weinan Zhang, Xialin He, Chen Yu, Zheng Tian, Yong Yu, Jun Wang
Deep reinforcement learning has recently emerged as an appealing alternative for legged locomotion over multiple terrains by training a policy in physical simulation and then transferring it to the real world (i. e., sim-to-real transfer).
no code implementations • 12 Dec 2022 • Lixin Cao, Jun Wang, Ben Yang, Dan Su, Dong Yu
Self-supervised learning (SSL) models confront challenges of abrupt informational collapse or slow dimensional collapse.
no code implementations • 8 Dec 2022 • Kaiyuan Tan, Jun Wang, Yiannis Kantaros
To bridge this gap, in this paper, we propose a targeted adversarial attack against DNN models for trajectory forecasting tasks.
1 code implementation • Neural Computing and Applications 2022 • Xianlin Peng, Huayu Zhao, Xiaoyu Wang, Yongqin Zhang, Zhan Li, Qunxi Zhang, Jun Wang, Jinye Peng, Haida Liang
Our network also uses dual-domain partial convolution with a mask for computing on only valid points, whereas the mask is updated for the next layer.
no code implementations • 5 Dec 2022 • Yourui Huangfu, Jian Wang, Shengchen Dai, Rong Li, Jun Wang, Chongwen Huang, Zhaoyang Zhang
The statistical data hinder the trained AI models from further fine-tuning for a specific scenario, and ray-tracing data with limited environments lower down the generalization capability of the trained AI models.
no code implementations • 28 Nov 2022 • Zijun Gao, Jun Wang, Guoxian Yu, Zhongmin Yan, Carlotta Domeniconi, Jinglin Zhang
LtCMH firstly adopts auto-encoders to mine the individuality and commonality of different modalities by minimizing the dependency between the individuality of respective modalities and by enhancing the commonality of these modalities.
no code implementations • 22 Nov 2022 • Dezhi Yang, Guoxian Yu, Jun Wang, Zhengtian Wu, Maozu Guo
In this paper, we propose {Reinforcement Causal Structure Learning on Order Graph} (RCL-OG) that uses order graph instead of MCMC to model different DAG topological orderings and to reduce the problem size.
no code implementations • 21 Nov 2022 • Junjie Sheng, Lu Wang, Fangkai Yang, Bo Qiao, Hang Dong, Xiangfeng Wang, Bo Jin, Jun Wang, Si Qin, Saravan Rajmohan, QIngwei Lin, Dongmei Zhang
To address these two limitations, this paper formulates the oversubscription for cloud as a chance-constrained optimization problem and propose an effective Chance Constrained Multi-Agent Reinforcement Learning (C2MARL) method to solve this problem.
Multi-agent Reinforcement Learning
reinforcement-learning
+1
no code implementations • 15 Nov 2022 • Runji Lin, Ye Li, Xidong Feng, Zhaowei Zhang, Xian Hong Wu Fung, Haifeng Zhang, Jun Wang, Yali Du, Yaodong Yang
Firstly, we propose prompt tuning for offline RL, where a context vector sequence is concatenated with the input to guide the conditional policy generation.
no code implementations • 2 Nov 2022 • Jun Wang, Abhir Bhalerao, Terry Yin, Simon See, Yulan He
Radiology report generation (RRG) has gained increasing research attention because of its huge potential to mitigate medical resource shortages and aid the process of disease decision making by radiologists.
no code implementations • 21 Oct 2022 • Jun Wang, Weixun Li, Changyu Hou, Xin Tang, Yixuan Qiao, Rui Fang, Pengyong Li, Peng Gao, Guotong Xie
Contrastive learning has emerged as a powerful tool for graph representation learning.
no code implementations • 18 Oct 2022 • Yangheng Zhao, Jun Wang, Xiaolong Li, Yue Hu, Ce Zhang, Yanfeng Wang, Siheng Chen
Instead of learning a single prototype for each class, in this paper, we propose to use an adaptive number of prototypes to dynamically describe the different point patterns within a semantic class.
Ranked #12 on
3D Semantic Segmentation
on SemanticKITTI
no code implementations • 15 Oct 2022 • Ziqing Wang, Zhirong Ye, Yuyang Du, Yi Mao, Yanying Liu, Ziling Wu, Jun Wang
DBSCAN has been widely used in density-based clustering algorithms.
1 code implementation • 14 Oct 2022 • Qianying Liu, Chaitanya Kaul, Jun Wang, Christos Anagnostopoulos, Roderick Murray-Smith, Fani Deligianni
For medical image semantic segmentation (MISS), Vision Transformers have emerged as strong alternatives to convolutional neural networks thanks to their inherent ability to capture long-range correlations.
no code implementations • 11 Oct 2022 • You Guo, Jun Wang, Trevor Cohn
Deep neural networks are vulnerable to adversarial attacks, such as backdoor attacks in which a malicious adversary compromises a model during training such that specific behaviour can be triggered at test time by attaching a specific word or phrase to an input.
1 code implementation • 30 Sep 2022 • Donghan Yu, Sheng Zhang, Patrick Ng, Henghui Zhu, Alexander Hanbo Li, Jun Wang, Yiqun Hu, William Wang, Zhiguo Wang, Bing Xiang
Question answering over knowledge bases (KBs) aims to answer natural language questions with factual information such as entities and relations in KBs.
no code implementations • 28 Sep 2022 • Jun Wang, Patrick Ng, Alexander Hanbo Li, Jiarong Jiang, Zhiguo Wang, Ramesh Nallapati, Bing Xiang, Sudipta Sengupta
When synthesizing a SQL query, there is no explicit semantic information of NLQ available to the parser which leads to undesirable generalization performance.
no code implementations • 22 Sep 2022 • Wuque Cai, Hongze Sun, Rui Liu, Yan Cui, Jun Wang, Yang Xia, Dezhong Yao, Daqing Guo
Spiking neural networks (SNNs) mimic brain computational strategies, and exhibit substantial capabilities in spatiotemporal information processing.
no code implementations • 19 Sep 2022 • Hailin Shi, Hang Du, Yibo Hu, Jun Wang, Dan Zeng, Ting Yao
Such multi-shot scheme brings inference burden, and the predefined scales inevitably have gap from real data.
1 code implementation • 17 Sep 2022 • Wei Liu, Haozhao Wang, Jun Wang, Ruixuan Li, Chao Yue, Yuankai Zhang
Conventional works generally employ a two-phase model in which a generator selects the most important pieces, followed by a predictor that makes predictions based on the selected pieces.
no code implementations • 14 Sep 2022 • Jun Wang
Extensive results on two well-known meeting datasets (AMI and ICSI corpora) show the effectiveness of our direct speech-based method to improve the summarization quality with untranscribed data.
no code implementations • 10 Sep 2022 • Alexander I. Cowen-Rivers, Philip John Gorinski, Aivar Sootla, Asif Khan, Liu Furui, Jun Wang, Jan Peters, Haitham Bou Ammar
Optimizing combinatorial structures is core to many real-world problems, such as those encountered in life sciences.
no code implementations • ACL 2022 • Xuemei Tang, Qi Su, Jun Wang
The evolution of language follows the rule of gradual change.
no code implementations • 6 Sep 2022 • Tianfu Qi, Jun Wang, Xiaonan Chen, Wei Huang
In many scenarios, the communication system suffers from both Gaussian white noise and non-Gaussian impulsive noise.
no code implementations • 5 Sep 2022 • Fei Hu, Honghua Chen, Xuequan Lu, Zhe Zhu, Jun Wang, Weiming Wang, Fu Lee Wang, Mingqiang Wei
We propose a novel stepwise point cloud completion network (SPCNet) for various 3D models with large missings.
no code implementations • 2 Sep 2022 • Taher Jafferjee, Juliusz Ziomek, Tianpei Yang, Zipeng Dai, Jianhong Wang, Matthew Taylor, Kun Shao, Jun Wang, David Mguni
Because MARL agents explore and update their policies during training, these observations often provide poor predictions about other agents' behaviour and the expected return for a given action.
Multi-agent Reinforcement Learning
reinforcement-learning
+2
1 code implementation • 2 Sep 2022 • Omran Alamayreh, Giovanna Maria Dimitri, Jun Wang, Benedetta Tondi, Mauro Barni
Notably, we found that asking the network to identify the country provides better results than estimating the geo-coordinates and then tracing them back to the country where the picture was taken.
no code implementations • 2 Sep 2022 • Honghua Chen, Mingqiang Wei, Jun Wang
In this work, we provide a comprehensive review of the advances in mesh denoising, containing both traditional geometric approaches and recent learning-based methods.
no code implementations • 30 Aug 2022 • Xintong Qin, Zhengyu Song, Tianwei Hou, Wenjuan Yu, Jun Wang, Xin Sun
The unmanned aerial vehicle (UAV) enabled mobile edge computing (MEC) has been deemed a promising paradigm to provide ubiquitous communication and computing services for the Internet of Things (IoT).
no code implementations • 30 Aug 2022 • Zhengyu Song, Xintong Qin, Yuanyuan Hao, Tianwei Hou, Jun Wang, Xin Sun
Driven by the visions of Internet of Things (IoT), there is an ever-increasing demand for computation resources of IoT users to support diverse applications.
1 code implementation • 30 Aug 2022 • Anyi Huang, Qian Xie, Zhoutao Wang, Dening Lu, Mingqiang Wei, Jun Wang
Second, a multi-scale perception module is designed to embed multi-scale geometric information for each scale feature and regress multi-scale weights to guide a multi-offset denoising displacement.
no code implementations • 30 Aug 2022 • Wei zhang, Zhaohong Deng, Kup-Sze Choi, Jun Wang, Shitong Wang
Meanwhile, to make the representation learning more specific to the clustering task, a one-step learning framework is proposed to integrate representation learning and clustering partition as a whole.
no code implementations • 24 Aug 2022 • Kai Liang, Jun Wang, Abhir Bhalerao
Previous works often adopt physical variables such as driving speed, acceleration and so forth for lane change classification.
1 code implementation • 21 Aug 2022 • Ashkan Farhangi, Jiang Bian, Arthur Huang, Haoyi Xiong, Jun Wang, Zhishan Guo
Moreover, the framework employs a dynamic uncertainty optimization algorithm that reduces the uncertainty of forecasts in an online manner.
no code implementations • 10 Aug 2022 • Xuxiang Jiang, Yinhao Xiao, Jun Wang, Wei zhang
Vulnerability identification is crucial for cyber security in the software-related industry.
no code implementations • 8 Aug 2022 • Xin Liu, Wei Tao, Jun Wang, Zhisong Pan
Due to the simplicity and efficiency of the first-order gradient method, it has been widely used in training neural networks.
1 code implementation • 4 Aug 2022 • Zhilei Chen, Honghua Chen, Lina Gong, Xuefeng Yan, Jun Wang, Yanwen Guo, Jing Qin, Mingqiang Wei
High-confidence overlap prediction and accurate correspondences are critical for cutting-edge models to align paired point clouds in a partial-to-partial manner.
1 code implementation • 3 Aug 2022 • Jun Wang, Mingfei Gao, Yuqian Hu, Ramprasaath R. Selvaraju, Chetan Ramaiah, ran Xu, Joseph F. JaJa, Larry S. Davis
To address this deficiency, we develop a new method to generate high-quality and diverse QA pairs by explicitly utilizing the existing rich text available in the scene context of each image.
1 code implementation • 2 Aug 2022 • Jakub Grudzien Kuba, Xidong Feng, Shiyao Ding, Hao Dong, Jun Wang, Yaodong Yang
The necessity for cooperation among intelligent machines has popularised cooperative multi-agent reinforcement learning (MARL) in the artificial intelligence (AI) research community.
no code implementations • 1 Aug 2022 • Zhe Zhu, Liangliang Nan, Haoran Xie, Honghua Chen, Mingqiang Wei, Jun Wang, Jing Qin
The first module transfers the intrinsic shape characteristics from single images to guide the geometry generation of the missing regions of point clouds, in which we propose IPAdaIN to embed the global features of both the image and the partial point cloud into completion.
Ranked #2 on
Point Cloud Completion
on ShapeNet-ViPC
no code implementations • 26 Jul 2022 • Zeren Huang, WenHao Chen, Weinan Zhang, Chuhan Shi, Furui Liu, Hui-Ling Zhen, Mingxuan Yuan, Jianye Hao, Yong Yu, Jun Wang
Deriving a good variable selection strategy in branch-and-bound is essential for the efficiency of modern mixed-integer programming (MIP) solvers.
2 code implementations • 13 Jul 2022 • Yali Du, Chengdong Ma, Yuchen Liu, Runji Lin, Hao Dong, Jun Wang, Yaodong Yang
Reinforcement learning algorithms require a large amount of samples; this often limits their real-world applications on even simple tasks.
1 code implementation • 11 Jul 2022 • Jun Wang, Abhir Bhalerao, Yulan He
Radiology report generation (RRG) aims to describe automatically a radiology image with human-like language and could potentially support the work of radiologists, reducing the burden of manual reporting.
1 code implementation • 7 Jul 2022 • Chengfeng Zhou, Songchang Chen, Chenming Xu, Jun Wang, Feng Liu, Chun Zhang, Juan Ye, Hefeng Huang, Dahong Qian
In this study, we present a novel normalization technique called window normalization (WIN) to improve the model generalization on heterogeneous medical images, which is a simple yet effective alternative to existing normalization methods.
no code implementations • 2 Jul 2022 • Honghua Chen, Zeyong Wei, Yabin Xu, Mingqiang Wei, Jun Wang
Low-overlap regions between paired point clouds make the captured features very low-confidence, leading cutting edge models to point cloud registration with poor quality.
1 code implementation • 9 Jun 2022 • Mingqiang Wei, Zeyong Wei, Haoran Zhou, Fei Hu, Huajian Si, Zhilei Chen, Zhe Zhu, Jingbo Qiu, Xuefeng Yan, Yanwen Guo, Jun Wang, Jing Qin
In this paper, we propose Adaptive Graph Convolution (AGConv) for wide applications of point cloud analysis.
1 code implementation • 6 Jun 2022 • Aivar Sootla, Alexander I. Cowen-Rivers, Jun Wang, Haitham Bou Ammar
We further show that Simmer can stabilize training and improve the performance of safe RL with average constraints.
no code implementations • 31 May 2022 • David Mguni, Aivar Sootla, Juliusz Ziomek, Oliver Slumbers, Zipeng Dai, Kun Shao, Jun Wang
In this paper, we introduce a reinforcement learning (RL) framework named \textbf{L}earnable \textbf{I}mpulse \textbf{C}ontrol \textbf{R}einforcement \textbf{A}lgorithm (LICRA), for learning to optimally select both when to act and which actions to take when actions incur costs.
no code implementations • 31 May 2022 • Jun Shi, Yuanming Zhang, Zheng Li, Xiangmin Han, Saisai Ding, Jun Wang, Shihui Ying
In this work, we propose a pseudo-data based self-supervised federated learning (FL) framework, named SSL-FT-BT, to improve both the diagnostic accuracy and generalization of CAD models.
no code implementations • 30 May 2022 • Oliver Slumbers, David Henry Mguni, Stephen Marcus McAleer, Stefano B. Blumberg, Jun Wang, Yaodong Yang
Although there are equilibrium concepts in game theory that take into account risk aversion, they either assume that agents are risk-neutral with respect to the uncertainty caused by the actions of other agents, or they are not guaranteed to exist.
no code implementations • 30 May 2022 • Changmin Yu, David Mguni, Dong Li, Aivar Sootla, Jun Wang, Neil Burgess
Efficient reinforcement learning (RL) involves a trade-off between "exploitative" actions that maximise expected reward and "explorative'" ones that sample unvisited states.
1 code implementation • 30 May 2022 • Muning Wen, Jakub Grudzien Kuba, Runji Lin, Weinan Zhang, Ying Wen, Jun Wang, Yaodong Yang
In this paper, we introduce a novel architecture named Multi-Agent Transformer (MAT) that effectively casts cooperative multi-agent reinforcement learning (MARL) into SM problems wherein the task is to map agents' observation sequence to agents' optimal action sequence.
no code implementations • SemEval (NAACL) 2022 • Changyu Hou, Jun Wang, Yixuan Qiao, Peng Jiang, Peng Gao, Guotong Xie, Qizhi Lin, Xiaopeng Wang, Xiandi Jiang, Benqi Wang, Qifeng Xiao
By assigning different weights to each model for different inputs, we adopted the Transformer layer to integrate the advantages of diverse models effectively.
Low Resource Named Entity Recognition
named-entity-recognition
+2
no code implementations • 27 May 2022 • Alexandre Maraval, Matthieu Zimmer, Antoine Grosnit, Rasul Tutunov, Jun Wang, Haitham Bou Ammar
First, we notice that these models are trained on uniformly distributed inputs, which impairs predictive accuracy on non-uniform data - a setting arising from any typical BO loop due to exploration-exploitation trade-offs.
no code implementations • 26 May 2022 • Jun Wang, Sixian Wang, Jincheng Dai, Zhongwei Si, Dekun Zhou, Kai Niu
However, current deep JSCC image transmission systems are typically optimized for traditional distortion metrics such as peak signal-to-noise ratio (PSNR) or multi-scale structural similarity (MS-SSIM).
1 code implementation • 22 May 2022 • Fanglei Sun, Yang Li, Ying Wen, Jingchen Hu, Jun Wang, Yang Yang, Kai Li
The design of MAFENN framework and algorithm are dedicated to enhance the learning capability of the feedfoward DL networks or their variations with the simple data feedback.
1 code implementation • 20 May 2022 • Shangding Gu, Long Yang, Yali Du, Guang Chen, Florian Walter, Jun Wang, Yaodong Yang, Alois Knoll
To establish a good foundation for future research in this thread, in this paper, we provide a review for safe RL from the perspectives of methods, theory and applications.
no code implementations • 18 May 2022 • Yixuan Qiao, Hao Chen, Jun Wang, Yongquan Lai, Tuozhen Liu, Xianbin Ye, Xin Tang, Rui Fang, Peng Gao, Wenfeng Xie, Guotong Xie
This paper describes the PASH participation in TREC 2021 Deep Learning Track.
no code implementations • 14 May 2022 • Jun Wang, Omran Alamayreh, Benedetta Tondi, Mauro Barni
In this paper, we address a new image forensics task, namely the detection of fake flood images generated by ClimateGAN architecture.
1 code implementation • ACL 2022 • Yang Li, Cheng Yu, Guangzhi Sun, Hua Jiang, Fanglei Sun, Weiqin Zu, Ying Wen, Yang Yang, Jun Wang
Modelling prosody variation is critical for synthesizing natural and expressive speech in end-to-end text-to-speech (TTS) systems.
no code implementations • 3 May 2022 • Yurong Chen, Xiaotie Deng, Chenchen Li, David Mguni, Jun Wang, Xiang Yan, Yaodong Yang
Fictitious play (FP) is one of the most fundamental game-theoretical learning frameworks for computing Nash equilibrium in $n$-player games, which builds the foundation for modern multi-agent learning algorithms.
no code implementations • 28 Apr 2022 • Guochang Lin, Fukai Chen, Pipi Hu, Xiang Chen, Junqing Chen, Jun Wang, Zuoqiang Shi
In addition, we also use the Green's function calculated by our method to solve a class of PDE, and also obtain high-precision solutions, which shows the good generalization ability of our method on solving PDEs.
no code implementations • 24 Apr 2022 • Wenbin Song, Mingrui Zhang, Joseph G. Wallwork, Junpeng Gao, Zheng Tian, Fanglei Sun, Matthew D. Piggott, Junqing Chen, Zuoqiang Shi, Xiang Chen, Jun Wang
However, mesh movement methods, such as the Monge-Ampere method, require the solution of auxiliary equations, which can be extremely expensive especially when the mesh is adapted frequently.
2 code implementations • 21 Apr 2022 • Rongjie Huang, Max W. Y. Lam, Jun Wang, Dan Su, Dong Yu, Yi Ren, Zhou Zhao
Also, FastDiff enables a sampling speed of 58x faster than real-time on a V100 GPU, making diffusion models practically applicable to speech synthesis deployment for the first time.
Ranked #7 on
Text-To-Speech Synthesis
on LJSpeech
(using extra training data)
no code implementations • ACL 2022 • Huibin Zhang, Zhengkun Zhang, Yao Zhang, Jun Wang, Yufan Li, Ning Jiang, Xin Wei, Zhenglu Yang
Procedural Multimodal Documents (PMDs) organize textual instructions and corresponding images step by step.
no code implementations • 2 Apr 2022 • Zeyong Wei, Honghua Chen, Hao Tang, Qian Xie, Mingqiang Wei, Jun Wang
The shape of circle is one of fundamental geometric primitives of man-made engineering objects.
1 code implementation • ICLR 2022 • Max W. Y. Lam, Jun Wang, Dan Su, Dong Yu
We propose a new bilateral denoising diffusion model (BDDM) that parameterizes both the forward and reverse processes with a schedule network and a score network, which can train with a novel bilateral modeling objective.
Ranked #1 on
Speech Synthesis
on LJSpeech
1 code implementation • 23 Mar 2022 • Haoran Zhou, Honghua Chen, Yingkui Zhang, Mingqiang Wei, Haoran Xie, Jun Wang, Tong Lu, Jing Qin, Xiao-Ping Zhang
Differently, our network is designed to refine the initial normal of each point by extracting additional information from multiple feature representations.
1 code implementation • 16 Mar 2022 • Jun Wang, Ying Cui, Dongyan Guo, Junxia Li, Qingshan Liu, Chunhua Shen
To solve the problems, we leverage the cross-attention and self-attention mechanisms to design novel neural network for processing point cloud in a per-point manner to eliminate kNNs.
2 code implementations • 4 Mar 2022 • Minghuan Liu, Zhengbang Zhu, Yuzheng Zhuang, Weinan Zhang, Jianye Hao, Yong Yu, Jun Wang
Recent progress in state-only imitation learning extends the scope of applicability of imitation learning to real-world settings by relieving the need for observing expert actions.
no code implementations • 2 Mar 2022 • Xianbin Ye, Ziliang Li, Fei Ma, Zongbi Yi, Pengyong Li, Jun Wang, Peng Gao, Yixuan Qiao, Guotong Xie
Anti-cancer drug discoveries have been serendipitous, we sought to present the Open Molecular Graph Learning Benchmark, named CandidateDrug4Cancer, a challenging and realistic benchmark dataset to facilitate scalable, robust, and reproducible graph machine learning research for anti-cancer drug discovery.
no code implementations • 18 Feb 2022 • Yiyi Zhang, Ying Zheng, Xiaogang Xu, Jun Wang
In this paper, we investigate the role of self-supervised representation learning in the context of CDFSL via a thorough evaluation of existing methods.
no code implementations • 17 Feb 2022 • Mengyue Yang, Xinyu Cai, Furui Liu, Xu Chen, Zhitang Chen, Jianye Hao, Jun Wang
It is evidence that representation learning can improve model's performance over multiple downstream tasks in many real-world scenarios, such as image classification and recommender systems.
1 code implementation • 14 Feb 2022 • Aivar Sootla, Alexander I. Cowen-Rivers, Taher Jafferjee, Ziyan Wang, David Mguni, Jun Wang, Haitham Bou-Ammar
Satisfying safety constraints almost surely (or with probability one) can be critical for the deployment of Reinforcement Learning (RL) in real-life applications.
no code implementations • ICLR 2022 • Hang Ren, Aivar Sootla, Taher Jafferjee, Junxiao Shen, Jun Wang, Haitham Bou-Ammar
We consider a context-dependent Reinforcement Learning (RL) setting, which is characterized by: a) an unknown finite number of not directly observable contexts; b) abrupt (discontinuous) context changes occurring during an episode; and c) Markovian context evolution.
no code implementations • 10 Feb 2022 • Juliusz Krysztof Ziomek, Jun Wang, Yaodong Yang
We study a novel setting in offline reinforcement learning (RL) where a number of distributed machines jointly cooperate to solve the problem but only one single round of communication is allowed and there is a budget constraint on the total number of information (in terms of bits) that each machine can send out.
no code implementations • 10 Feb 2022 • Yingzhanghao Zhou, Xiang Chen, Peng Zhang, Jun Wang, Lei Wang, Hong Guo
Since proposed in the 70s, the Non-Equilibrium Green Function (NEGF) method has been recognized as a standard approach to quantum transport simulations.
1 code implementation • 9 Feb 2022 • Moyi Yang, Junjie Sheng, Xiangfeng Wang, Wenyan Liu, Bo Jin, Jun Wang, Hongyuan Zha
Fairness has been taken as a critical metric in machine learning models, which is considered as an important component of trustworthy machine learning.
no code implementations • 3 Feb 2022 • Xihan Li, Xiang Chen, Rasul Tutunov, Haitham Bou-Ammar, Lei Wang, Jun Wang
The Schr\"odinger equation is at the heart of modern quantum mechanics.
1 code implementation • 3 Feb 2022 • Yabin Xu, Liangliang Nan, Laishui Zhou, Jun Wang, Charlie C. L. Wang
However, due to the discrete nature and limited resolution of their surface representations (e. g., point- or voxel-based), existing approaches suffer from the accumulation of errors in camera tracking and distortion in the reconstruction, which leads to an unsatisfactory 3D reconstruction.
1 code implementation • 29 Jan 2022 • Asif Khan, Alexander I. Cowen-Rivers, Antoine Grosnit, Derrick-Goh-Xin Deik, Philippe A. Robert, Victor Greiff, Eva Smorodina, Puneet Rawat, Kamil Dreczkowski, Rahmad Akbar, Rasul Tutunov, Dany Bou-Ammar, Jun Wang, Amos Storkey, Haitham Bou-Ammar
software suite as a black-box oracle to score the target specificity and affinity of designed antibodies \textit{in silico} in an unconstrained fashion~\citep{robert2021one}.
no code implementations • 28 Jan 2022 • Ming Zhou, Jingxiao Chen, Ying Wen, Weinan Zhang, Yaodong Yang, Yong Yu, Jun Wang
Policy Space Response Oracle methods (PSRO) provide a general solution to learn Nash equilibrium in two-player zero-sum games but suffer from two drawbacks: (1) the computation inefficiency due to the need for consistent meta-game evaluation via simulations, and (2) the exploration inefficiency due to finding the best response against a fixed meta-strategy at every epoch.
no code implementations • 25 Jan 2022 • Jianwei Xu, Ran Zhao, Yizhou Yu, Qingwei Zhang, Xianzhang Bian, Jun Wang, Zhizheng Ge, Dahong Qian
In order to solve these problems, our method combines the two-dimensional (2-D) CNN-based real-time object detector network with spatiotemporal information.
no code implementations • 24 Jan 2022 • Wei Huang, Tianfu Qi, Yundi Guan, Qihang Peng, Jun Wang
In this paper, we investigate the waveform design from a novel perspective and propose a new waveform design paradigm with the knowledge graph (KG)-based intelligent recommendation system.
no code implementations • 22 Jan 2022 • Xuemei Tang, Jun Wang, Qi Su
In recent years, deep learning has achieved significant success in the Chinese word segmentation (CWS) task.
no code implementations • 16 Jan 2022 • Mengyue Yang, Guohao Cai, Furui Liu, Zhenhua Dong, Xiuqiang He, Jianye Hao, Jun Wang, Xu Chen
To alleviate these problems, in this paper, we propose a novel debiased recommendation framework based on user feature balancing.
1 code implementation • 12 Jan 2022 • Xue Yan, Yali Du, Binxin Ru, Jun Wang, Haifeng Zhang, Xu Chen
The Elo rating system is widely adopted to evaluate the skills of (chess) game and sports players.
no code implementations • 10 Jan 2022 • Yangyang Wu, Jun Wang, Xiaoye Miao, Wenjia Wang, Jianwei Yin
DIM leverages a new masking Sinkhorn divergence function to make an arbitrary generative adversarial imputation model differentiable, while for such a differentiable imputation model, SSE can estimate an appropriate sample size to ensure the user-specified imputation accuracy of the final model.
no code implementations • 31 Dec 2021 • Yuang Liu, Wei zhang, Jun Wang, Jianyong Wang
In this paper, we provide a comprehensive survey on data-free knowledge transfer from the perspectives of knowledge distillation and unsupervised domain adaptation, to help readers have a better understanding of the current research status and ideas.
1 code implementation • 31 Dec 2021 • Bo Liu, Xidong Feng, Jie Ren, Luo Mai, Rui Zhu, Haifeng Zhang, Jun Wang, Yaodong Yang
Gradient-based Meta-RL (GMRL) refers to methods that maintain two-level optimisation procedures wherein the outer-loop meta-learner guides the inner-loop gradient-based reinforcement learner to achieve fast adaptations.
no code implementations • 9 Dec 2021 • Jun Wang, Zhoujing Li, Yixuan Qiao, Qiming Qin, Peng Gao, Guotong Xie
This paper presents a novel superpixel based approach combining DNN and a modified segmentation method, to detect damaged buildings from VHR imagery.
no code implementations • 7 Dec 2021 • Tianyang Hu, Jun Wang, Wenjia Wang, Zhenguo Li
Comparing to cross-entropy, square loss has comparable generalization error but noticeable advantages in robustness and model calibration.
1 code implementation • 6 Dec 2021 • Linghui Meng, Muning Wen, Yaodong Yang, Chenyang Le, Xiyun Li, Weinan Zhang, Ying Wen, Haifeng Zhang, Jun Wang, Bo Xu
In this paper, we facilitate the research by providing large-scale datasets, and use them to examine the usage of the Decision Transformer in the context of MARL.
1 code implementation • NeurIPS 2021 • Xidong Feng, Oliver Slumbers, Ziyu Wan, Bo Liu, Stephen Mcaleer, Ying Wen, Jun Wang, Yaodong Yang
When solving two-player zero-sum games, multi-agent reinforcement learning (MARL) algorithms often create populations of agents where, at each iteration, a new agent is discovered as the best response to a mixture over the opponent population.
Multi-agent Reinforcement Learning
Vocal Bursts Valence Prediction
1 code implementation • ICLR 2022 • Changmin Yu, Dong Li, Jianye Hao, Jun Wang, Neil Burgess
We propose learning via retracing, a novel self-supervised approach for learning the state representation (and the associated dynamics model) for reinforcement learning tasks.
no code implementations • 11 Nov 2021 • Antoine Grosnit, Cedric Malherbe, Rasul Tutunov, Xingchen Wan, Jun Wang, Haitham Bou Ammar
Optimising the quality-of-results (QoR) of circuits during logic synthesis is a formidable challenge necessitating the exploration of exponentially sized search spaces.
no code implementations • 7 Nov 2021 • Yuanlin Yang, Guoxian Yu, Jun Wang, Lei Liu, Carlotta Domeniconi, Maozu Guo
Multi-Instance Multi-Label learning (MIML) models complex objects (bags), each of which is associated with a set of interrelated labels and composed with a set of instances.
no code implementations • 28 Oct 2021 • Yunxiang Li, Jingxiong Li, Ruilong Dan, Shuai Wang, Kai Jin, Guodong Zeng, Jun Wang, Xiangji Pan, Qianni Zhang, Huiyu Zhou, Qun Jin, Li Wang, Yaqi Wang
To mitigate this problem, a novel unsupervised domain adaptation (UDA) method named dispensed Transformer network (DTNet) is introduced in this paper.
no code implementations • 28 Oct 2021 • Chenguang Wang, Yaodong Yang, Oliver Slumbers, Congying Han, Tiande Guo, Haifeng Zhang, Jun Wang
In this paper, we introduce a two-player zero-sum framework between a trainable \emph{Solver} and a \emph{Data Generator} to improve the generalization ability of deep learning-based solvers for Traveling Salesman Problem (TSP).
no code implementations • 27 Oct 2021 • David Mguni, Usman Islam, Yaqi Sun, Xiuling Zhang, Joel Jennings, Aivar Sootla, Changmin Yu, Ziyan Wang, Jun Wang, Yaodong Yang
In this paper, we introduce a new generation of RL solvers that learn to minimise safety violations while maximising the task reward to the extent that can be tolerated by the safe policy.
no code implementations • 26 Oct 2021 • Pengyong Li, Jun Wang, Ziliang Li, Yixuan Qiao, Xianggen Liu, Fei Ma, Peng Gao, Seng Song, Guotong Xie
Self-supervised learning has gradually emerged as a powerful technique for graph representation learning.
no code implementations • 22 Oct 2021 • Ricky Sanjaya, Jun Wang, Yaodong Yang
In this paper, we quantify the non-transitivity in Chess through real-world data from human players.
no code implementations • 21 Oct 2021 • Yanbin He, Zhiyang Lu, Jun Wang, Jun Shi
Convolutional neural networks (CNNs) and their variants have been successfully applied to the electroencephalogram (EEG) based motor imagery (MI) decoding task.
no code implementations • 7 Oct 2021 • Le Cong Dinh, David Henry Mguni, Long Tran-Thanh, Jun Wang, Yaodong Yang
In this setting, we first demonstrate that MDP-Expert, an existing algorithm that works well with oblivious adversaries can still apply and achieve a policy regret bound of $\mathcal{O}(\sqrt{T \log(L)}+\tau^2\sqrt{ T \log(|A|)})$ where $L$ is the size of adversary's pure strategy set and $|A|$ denotes the size of agent's action space.
3 code implementations • 6 Oct 2021 • Shangding Gu, Jakub Grudzien Kuba, Munning Wen, Ruiqing Chen, Ziyan Wang, Zheng Tian, Jun Wang, Alois Knoll, Yaodong Yang
To fill these gaps, in this work, we formulate the safe MARL problem as a constrained Markov game and solve it with policy optimisation methods.
Multi-agent Reinforcement Learning
reinforcement-learning
+1
1 code implementation • 30 Sep 2021 • Yunxiang Li, Shuai Wang, Jun Wang, Guodong Zeng, Wenjun Liu, Qianni Zhang, Qun Jin, Yaqi Wang
In this paper, we propose a novel end-to-end U-Net like Group Transformer Network (GT U-Net) for the tooth root segmentation.
no code implementations • 29 Sep 2021 • Jun Wang, Yiannis Kantaros
To mitigate this challenge, in this paper, we propose model-based robust adaptive training algorithm (MRTAdapt), a new training algorithm to enhance the robustness of DNN-based semantic segmentation methods against natural variations that leverages model-based robust training algorithms and generative adversarial networks.
no code implementations • 29 Sep 2021 • Hangyu Mao, Jianye Hao, Dong Li, Jun Wang, Weixun Wang, Xiaotian Hao, Bin Wang, Kun Shao, Zhen Xiao, Wulong Liu
In contrast, we formulate an \emph{explicit} credit assignment problem where each agent gives its suggestion about how to weight individual Q-values to explicitly maximize the joint Q-value, besides guaranteeing the Bellman optimality of the joint Q-value.
no code implementations • NeurIPS 2021 • Minghuan Liu, Zhengbang Zhu, Yuzheng Zhuang, Weinan Zhang, Jian Shen, Jianye Hao, Yong Yu, Jun Wang
State-only imitation learning (SOIL) enables agents to learn from massive demonstrations without explicit action or reward information.
no code implementations • 29 Sep 2021 • Jing Zhang, Peng Zhang, Yupeng He, Siwei Rao, Jun Wang, Guangjian Tian
In this framework, we derive the mathematical representation of the variable space, and then use a tensor network based on the idea of low-rank approximation to model the variable space.
no code implementations • 29 Sep 2021 • Jun Wang, Tianxiang Su
In our experiments, we compared the proposed method with multiple state-of-the-art continual learning methods and the mREMIND network outperformed others both in accuracy and in disk space usage.
no code implementations • 29 Sep 2021 • Guochang Lin, Fukai Chen, Pipi Hu, Xiang Chen, Junqing Chen, Jun Wang, Zuoqiang Shi
Green's function plays a significant role in both theoretical analysis and numerical computing of partial differential equations (PDEs).
no code implementations • 29 Sep 2021 • Linghui Meng, Muning Wen, Yaodong Yang, Chenyang Le, Xi yun Li, Haifeng Zhang, Ying Wen, Weinan Zhang, Jun Wang, Bo Xu
Offline reinforcement learning leverages static datasets to learn optimal policies with no necessity to access the environment.
Multi-agent Reinforcement Learning
reinforcement-learning
+2
no code implementations • 29 Sep 2021 • Yuchen Liu, Yali Du, Runji Lin, Hangrui Bi, Mingdong Wu, Jun Wang, Hao Dong
Model-based RL is an effective approach for reducing sample complexity.
Model-based Reinforcement Learning
Reinforcement Learning (RL)
no code implementations • 29 Sep 2021 • Rongjie Huang, Max W. Y. Lam, Jun Wang, Dan Su, Dong Yu, Zhou Zhao, Yi Ren
Learning generalizable speech representations for unseen samples in different domains has been a challenge with ever increasing importance to date.
no code implementations • 29 Sep 2021 • Mengyue Yang, Furui Liu, Xu Chen, Zhitang Chen, Jianye Hao, Jun Wang
In many real-world scenarios, such as image classification and recommender systems, it is evidence that representation learning can improve model's performance over multiple downstream tasks.
no code implementations • NeurIPS Workshop DLDE 2021 • Feng Zhao, Xiang Chen, Jun Wang, Zuoqiang Shi, Shao-Lun Huang
Traditionally, we provide technical parameters for ODE solvers, such as the order, the stepsize and the local error threshold.
5 code implementations • ICLR 2022 • Jakub Grudzien Kuba, Ruiqing Chen, Muning Wen, Ying Wen, Fanglei Sun, Jun Wang, Yaodong Yang
In this paper, we extend the theory of trust region learning to MARL.
1 code implementation • 23 Sep 2021 • Zhengwen Shen, Jun Wang, Zaiyu Pan, Yulian Li, Jiangyu Wang
In this paper, we propose a novel cross-attention-guided image fusion network, which is a unified and unsupervised framework for multi-modal image fusion, multi-exposure image fusion, and multi-focus image fusion.
no code implementations • 20 Sep 2021 • Yixin Wu, Rui Luo, Chen Zhang, Jun Wang, Yaodong Yang
In this paper, we characterize the noise of stochastic gradients and analyze the noise-induced dynamics during training deep neural networks by gradient-based optimizers.
no code implementations • 5 Sep 2021 • Dell Zhang, Jun Wang
Driven by the need to capture users' evolving interests and optimize their long-term experiences, more and more recommender systems have started to model recommendation as a Markov decision process and employ reinforcement learning to address the problem.
no code implementations • 4 Sep 2021 • Xiaotie Deng, Ningyuan Li, David Mguni, Jun Wang, Yaodong Yang
Similar to the role of Markov decision processes in reinforcement learning, Stochastic Games (SGs) lay the foundation for the study of multi-agent reinforcement learning (MARL) and sequential agent interactions.
Multi-agent Reinforcement Learning
reinforcement-learning
+1
no code implementations • 2 Sep 2021 • Mengyue Yang, Quanyu Dai, Zhenhua Dong, Xu Chen, Xiuqiang He, Jun Wang
To alleviate this problem, in this paper, we propose to reformulate the recommendation task within the causal inference framework, which enables us to counterfactually simulate user ranking-based preferences to handle the data scarce problem.
no code implementations • 26 Aug 2021 • Max W. Y. Lam, Jun Wang, Rongjie Huang, Dan Su, Dong Yu
In this paper, we propose novel bilateral denoising diffusion models (BDDMs), which take significantly fewer steps to generate high-quality samples.
no code implementations • 25 Aug 2021 • Jun Wang, Hefeng Zhou, Xiaohan Yu
There are two main problems hindering the performance of the two-phase WSOD approaches, i. e., insufficient learning problem and strict reliance between the FSD and the pseudo ground truth (PGT) generated by the WSOD model.
1 code implementation • 24 Aug 2021 • Xidong Feng, Chen Chen, Dong Li, Mengchen Zhao, Jianye Hao, Jun Wang
Meta learning, especially gradient based one, can be adopted to tackle this problem by learning initial parameters of the model and thus allowing fast adaptation to a specific task from limited data examples.
1 code implementation • NeurIPS 2021 • Jakub Grudzien Kuba, Muning Wen, Yaodong Yang, Linghui Meng, Shangding Gu, Haifeng Zhang, David Henry Mguni, Jun Wang
In multi-agent RL (MARL), although the PG theorem can be naturally extended, the effectiveness of multi-agent PG (MAPG) methods degrades as the variance of gradient estimates increases rapidly with the number of agents.