1 code implementation • ICML 2020 • Hangbo Bao, Li Dong, Furu Wei, Wenhui Wang, Nan Yang, Xiaodong Liu, Yu Wang, Jianfeng Gao, Songhao Piao, Ming Zhou, Hsiao-Wuen Hon
We propose to pre-train a unified language model for both autoencoding and partially autoregressive language modeling tasks using a novel training procedure, referred to as a pseudo-masked language model (PMLM).
no code implementations • 20 Mar 2023 • Nan Yang, Xuanyu Chen, Charles Z. Liu, Dong Yuan, Wei Bao, Lizhen Cui
Latest federated learning (FL) methods started to focus on how to use unlabeled data in clients for training due to users' privacy concerns, high labeling costs, or lack of expertise.
no code implementations • 14 Mar 2023 • Liang Wang, Nan Yang, Furu Wei
This paper introduces a simple yet effective query expansion approach, denoted as query2doc, to improve both sparse and dense retrieval systems.
no code implementations • 9 Mar 2023 • Shaohuai Liu, Jinbo Liu, Weirui Ye, Nan Yang, Guanglun Zhang, Haiwang Zhong, Chongqing Kang, Qirong Jiang, Xuri Song, Fangchun Di, Yang Gao
The well-trained scheduling agent significantly reduces renewable curtailment and load shedding, which are issues arising from traditional scheduling's reliance on inaccurate day-ahead forecasts.
no code implementations • 23 Feb 2023 • Nan Yang, Dong Yuan, Charles Z Liu, Yongkun Deng, Wei Bao
Most existing federated learning methods assume that clients have fully labeled data to train on, while in reality, it is hard for the clients to get task-specific labels due to users' privacy concerns, high labeling costs, or lack of expertise.
no code implementations • 17 Feb 2023 • Nan Yang, Laicheng Zhong, Fan Huang, Dong Yuan, Wei Bao
Random Padding is parameter-free, simple to construct, and compatible with the majority of CNN-based recognition models.
1 code implementation • 18 Jan 2023 • Felix Wimbauer, Nan Yang, Christian Rupprecht, Daniel Cremers
Currently, neural radiance fields (NeRFs) can capture true 3D including color but are too complex to be generated from a single image.
no code implementations • 31 Dec 2022 • Patrick Wenzel, Nan Yang, Rui Wang, Niclas Zeller, Daniel Cremers
In this paper, we present a novel visual SLAM and long-term localization benchmark for autonomous driving in challenging conditions based on the large-scale 4Seasons dataset.
1 code implementation • 7 Dec 2022 • Liang Wang, Nan Yang, Xiaolong Huang, Binxing Jiao, Linjun Yang, Daxin Jiang, Rangan Majumder, Furu Wei
This paper presents E5, a family of state-of-the-art text embeddings that transfer well to a wide range of tasks.
no code implementations • 2 Nov 2022 • Haolin Deng, Yanan Zhang, Yangfan Zhang, Wangyang Ying, Changlong Yu, Jun Gao, Wei Wang, Xiaoling Bai, Nan Yang, Jin Ma, Xiang Chen, Tianhua Zhou
To the best of our knowledge, it is currently the largest manually-annotated Chinese dataset for open event extraction.
no code implementations • 25 Oct 2022 • Nan Yang, Akram Shafie
Terahertz (THz) communications (THzCom) has experienced a meteoric rise of interest, due to its benefits for ultra-high data rate transmission in the sixth generation (6G) and beyond era.
no code implementations • 21 Oct 2022 • Zeyu Tang, Nan Yang, Simon Walsh, Guang Yang
Discontinuity in the delineation of peripheral bronchioles hinders the potential clinical application of automated airway segmentation models.
no code implementations • 22 Sep 2022 • Nan Yang, Xin Luan, Huidi Jia, Zhi Han, Yandong Tang
In this work, we put forward three concepts and corresponding definitions: editing continuity, consistency, and reversibility.
1 code implementation • 8 Aug 2022 • Zehan Li, Nan Yang, Liang Wang, Furu Wei
In this paper, we propose a new dense retrieval model which learns diverse document representations with deep query interactions.
no code implementations • 7 Aug 2022 • Akram Shafie, Chunhui Li, Nan Yang, Xiangyun Zhou, Trung Q. Duong
Numerical results demonstrate that comparing to existing approaches, our proposed unsupervised learning-based approach achieves a higher data rate, especially when the molecular absorption coefficient within the spectrum of interest varies in a highly non-linear manner.
no code implementations • 22 Jul 2022 • Akram Shafie, Nan Yang, Chong Han, Josep Miquel Jornet, Markku Juntti, Thomas Kurner
The unprecedented increase in wireless data traffic, predicted to occur within the next decade, is motivating academia and industries to look beyond contemporary wireless standards and conceptualize the sixth-generation (6G) wireless networks.
no code implementations • 6 Jul 2022 • Akram Shafie, Nan Yang, Chong Han, Josep M. Jornet
We also show that a further data rate gain can be obtained by optimally determining the unused spectra at the edges of TWs, as compared to avoiding using pre-defined spectra at the edges of TWs.
no code implementations • 6 Jul 2022 • Liang Wang, Nan Yang, Xiaolong Huang, Binxing Jiao, Linjun Yang, Daxin Jiang, Rangan Majumder, Furu Wei
In this paper, we propose SimLM (Similarity matching with Language Model pre-training), a simple yet effective pre-training method for dense passage retrieval.
1 code implementation • 10 Apr 2022 • Xinhang Li, Zihao Li, Nan Yang, Zheng Yuan, Qinwen Wang, Yiying Yang, Yupeng Huang, Xuri Song, Lei LI, Lin Zhang
The expansion of renewable energy could help realizing the goals of peaking carbon dioxide emissions and carbon neutralization.
no code implementations • 30 Mar 2022 • Tarun Yenamandra, Ayush Tewari, Nan Yang, Florian Bernard, Christian Theobalt, Daniel Cremers
To this end, we learn a signed distance function (SDF) along with our DDF model to represent a class of shapes.
1 code implementation • 14 Nov 2021 • Lukas Koestler, Nan Yang, Niclas Zeller, Daniel Cremers
In this paper, we present TANDEM a real-time monocular tracking and dense mapping framework.
no code implementations • 10 Nov 2021 • Akram Shafie, Nan Yang, Sheeraz Alvi, Chong Han, Salman Durrani, Josep M. Jornet
Aided by numerical results, we show that by enabling and optimizing ASB, significantly higher throughput can be achieved as compared to adopting equal sub-band bandwidth, and this throughput gain is most profound when the power budget constraint is more stringent.
1 code implementation • 26 Oct 2021 • Hangbo Bao, Li Dong, Wenhui Wang, Nan Yang, Furu Wei
Pretrained bidirectional Transformers, such as BERT, have achieved significant improvements in a wide variety of language understanding tasks, while it is not straightforward to directly apply them for natural language generation.
no code implementations • 3 Aug 2021 • Naili Xing, Sai Ho Yeung, ChengHao Cai, Teck Khim Ng, Wei Wang, Kaiyuan Yang, Nan Yang, Meihui Zhang, Gang Chen, Beng Chin Ooi
Specifically, in terms of usability, it is demanding for non-experts to implement deep learning models, obtain the right settings for the entire machine learning pipeline, manage models and datasets, and exploit external data sources all together.
no code implementations • ACL 2021 • Nan Yang, Furu Wei, Binxing Jiao, Daxing Jiang, Linjun Yang
Dense passage retrieval has been shown to be an effective approach for information retrieval tasks such as open domain question answering.
no code implementations • 20 Apr 2021 • Akram Shafie, Nan Yang, Salman Durrani, Xiangyun Zhou, Chong Han, Markku Juntti
We conduct novel coverage probability analysis of downlink transmission in a three-dimensional (3D) terahertz (THz) communication (THzCom) system.
1 code implementation • CVPR 2021 • Felix Wimbauer, Nan Yang, Lukas von Stumberg, Niclas Zeller, Daniel Cremers
Unlike other multi-view stereo methods, MonoRec is able to reconstruct both static and moving objects by leveraging the predicted masks.
no code implementations • 13 Oct 2020 • Lukas von Stumberg, Patrick Wenzel, Nan Yang, Daniel Cremers
The learned features significantly improve the robustness of direct image alignment, especially for relocalization across different conditions.
no code implementations • 14 Sep 2020 • Patrick Wenzel, Rui Wang, Nan Yang, Qing Cheng, Qadeer Khan, Lukas von Stumberg, Niclas Zeller, Daniel Cremers
We present a novel dataset covering seasonal and challenging perceptual conditions for autonomous driving.
4 code implementations • NAACL 2021 • Zewen Chi, Li Dong, Furu Wei, Nan Yang, Saksham Singhal, Wenhui Wang, Xia Song, Xian-Ling Mao, He-Yan Huang, Ming Zhou
In this work, we present an information-theoretic framework that formulates cross-lingual language model pre-training as maximizing mutual information between multilingual-multi-granularity texts.
Ranked #16 on
Zero-Shot Cross-Lingual Transfer
on XTREME
no code implementations • 7 Jul 2020 • Liangling Lai, Jinsong Hu, Youjia Chen, Haifeng Zheng, Nan Yang
We propose a new secure transmission scheme which uses directional modulation (DM) with artificial noise and is aided by the intelligent reflecting surface (IRS).
no code implementations • 16 Apr 2020 • Akram Shafie, Nan Yang, Zhuo Sun, Salman Durrani
We further show that the coverage performance improvement brought by increasing the antenna directivity at APs is higher than that brought by increasing the antenna directivity at UEs.
no code implementations • CVPR 2020 • Nan Yang, Lukas von Stumberg, Rui Wang, Daniel Cremers
We propose D3VO as a novel framework for monocular visual odometry that exploits deep networks on three levels -- deep depth, pose and uncertainty estimation.
3 code implementations • 28 Feb 2020 • Hangbo Bao, Li Dong, Furu Wei, Wenhui Wang, Nan Yang, Xiaodong Liu, Yu Wang, Songhao Piao, Jianfeng Gao, Ming Zhou, Hsiao-Wuen Hon
We propose to pre-train a unified language model for both autoencoding and partially autoregressive language modeling tasks using a novel training procedure, referred to as a pseudo-masked language model (PMLM).
Ranked #4 on
Question Generation
on SQuAD1.1
(using extra training data)
1 code implementation • NeurIPS 2020 • Wenhui Wang, Furu Wei, Li Dong, Hangbo Bao, Nan Yang, Ming Zhou
The small model (student) is trained by deeply mimicking the self-attention module, which plays a vital role in Transformer networks, of the large model (teacher).
Ranked #8 on
Zero-shot Text Search
on BEIR
no code implementations • WS 2019 • Hangbo Bao, Li Dong, Furu Wei, Wenhui Wang, Nan Yang, Lei Cui, Songhao Piao, Ming Zhou
Most machine reading comprehension (MRC) models separately handle encoding and matching with different network architectures.
no code implementations • 15 Oct 2019 • Eunah Jung, Nan Yang, Daniel Cremers
We propose the concept of a multi-frame GAN (MFGAN) and demonstrate its potential as an image sequence enhancement for stereo visual odometry in low light conditions.
no code implementations • 14 Oct 2019 • Hang Yuan, Nan Yang, Kai Yang, Chong Han, Jianping An
We consider a three-dimensional wideband THz channel by incorporating the joint effect of molecular absorption, high sparsity, and multi-path fading, and consider the carrier frequency offset in multi-carrier systems.
9 code implementations • NeurIPS 2019 • Li Dong, Nan Yang, Wenhui Wang, Furu Wei, Xiaodong Liu, Yu Wang, Jianfeng Gao, Ming Zhou, Hsiao-Wuen Hon
This paper presents a new Unified pre-trained Language Model (UniLM) that can be fine-tuned for both natural language understanding and generation tasks.
Ranked #2 on
Generative Question Answering
on CoQA
(using extra training data)
no code implementations • 22 Apr 2019 • Rui Wang, Nan Yang, Joerg Stueckler, Daniel Cremers
Scene understanding from images is a challenging problem encountered in autonomous driving.
no code implementations • 1 Dec 2018 • Yuting Fang, Adam Noel, Andrew W. Eckford, Nan Yang
The number of molecules observed at each randomly-distributed bacterium is first derived by characterizing the diffusion and degradation of signaling molecules within the population.
no code implementations • 17 Oct 2018 • Jianguo Chen, Kenli Li, Huigui Rong, Kashif Bilal, Nan Yang, Keqin Li
It is crucial to provide compatible treatment schemes for a disease according to various symptoms at different stages.
no code implementations • EMNLP 2018 • Minghao Hu, Yuxing Peng, Furu Wei, Zhen Huang, Dongsheng Li, Nan Yang, Ming Zhou
Despite that current reading comprehension systems have achieved significant advancements, their promising performances are often obtained at the cost of making an ensemble of numerous models.
no code implementations • 17 Aug 2018 • Minghao Hu, Furu Wei, Yuxing Peng, Zhen Huang, Nan Yang, Dongsheng Li
Machine reading comprehension with unanswerable questions aims to abstain from answering when no answer can be inferred.
Ranked #11 on
Question Answering
on SQuAD2.0 dev
1 code implementation • 6 Jul 2018 • Qingyu Zhou, Nan Yang, Furu Wei, Ming Zhou
Copying mechanism shows effectiveness in sequence-to-sequence based neural network models for text generation tasks, such as abstractive sentence summarization and question generation.
1 code implementation • ACL 2018 • Qingyu Zhou, Nan Yang, Furu Wei, Shaohan Huang, Ming Zhou, Tiejun Zhao
In this paper, we present a novel end-to-end neural network framework for extractive document summarization by jointly learning to score and select sentences.
Ranked #9 on
Extractive Text Summarization
on CNN / Daily Mail
no code implementations • ECCV 2018 • Nan Yang, Rui Wang, Jörg Stückler, Daniel Cremers
To this end, we incorporate deep depth predictions into Direct Sparse Odometry (DSO) as direct virtual stereo measurements.
no code implementations • ACL 2017 • Wenhui Wang, Nan Yang, Furu Wei, Baobao Chang, Ming Zhou
We first match the question and passage with gated attention-based recurrent networks to obtain the question-aware passage representation.
Ranked #35 on
Question Answering
on SQuAD1.1 dev
no code implementations • ACL 2017 • Shuangzhi Wu, Dong-dong Zhang, Nan Yang, Mu Li, Ming Zhou
Nowadays a typical Neural Machine Translation (NMT) model generates translations from left to right as a linear sequence, during which latent syntactic structures of the target sentences are not explicitly concerned.
no code implementations • 15 Jun 2017 • Chuanqi Tan, Furu Wei, Nan Yang, Bowen Du, Weifeng Lv, Ming Zhou
We build the answer extraction model with state-of-the-art neural networks for single passage reading comprehension, and propose an additional task of passage ranking to help answer extraction in multiple passages.
no code implementations • 23 May 2017 • Xin Guo, Johnny Hong, Nan Yang
Construction of ambiguity set in robust optimization relies on the choice of divergences between probability distributions.
no code implementations • 19 May 2017 • Xin Guo, Johnny Hong, Tianyi Lin, Nan Yang
Wasserstein Generative Adversarial Networks (WGANs) provide a versatile class of models, which have attracted great attention in various applications.
no code implementations • 11 May 2017 • Nan Yang, Rui Wang, Xiang Gao, Daniel Cremers
Monocular visual odometry (VO) and simultaneous localization and mapping (SLAM) have seen tremendous improvements in accuracy, robustness and efficiency, and have gained increasing popularity over recent years.
Monocular Visual Odometry
Simultaneous Localization and Mapping
2 code implementations • ACL 2017 • Qingyu Zhou, Nan Yang, Furu Wei, Ming Zhou
We propose a selective encoding model to extend the sequence-to-sequence framework for abstractive sentence summarization.
Ranked #8 on
Text Summarization
on DUC 2004 Task 1
4 code implementations • 6 Apr 2017 • Qingyu Zhou, Nan Yang, Furu Wei, Chuanqi Tan, Hangbo Bao, Ming Zhou
Automatic question generation aims to generate questions from a text passage where the generated questions can be answered by certain sub-spans of the given passage.
Ranked #13 on
Question Generation
on SQuAD1.1
no code implementations • COLING 2016 • Shi Feng, Shujie Liu, Nan Yang, Mu Li, Ming Zhou, Kenny Q. Zhu
In neural machine translation, the attention mechanism facilitates the translation process by producing a soft alignment between the source sentence and the target sentence.
no code implementations • 7 Dec 2015 • Bei Chen, Jun Zhu, Nan Yang, Tian Tian, Ming Zhou, Bo Zhang
Modeling document structure is of great importance for discourse analysis and related applications.
no code implementations • 18 Apr 2014 • Yaming Sun, Lei Lin, Duyu Tang, Nan Yang, Zhenzhou Ji, Xiaolong Wang
We present a method to leverage radical for learning Chinese character embedding.
no code implementations • 26 Nov 2013 • Vašek Chvátal, Mark Goldsmith, Nan Yang
In a pioneering classic, Warren McCulloch and Walter Pitts proposed a model of the central nervous system; motivated by EEG recordings of normal brain activity, Chv\' atal and Goldsmith asked whether or not this model can be engineered to provide pseudorandom number generators.