Search Results for author: Xilun Chen

Found 16 papers, 10 papers with code

Simple Local Attentions Remain Competitive for Long-Context Tasks

1 code implementation14 Dec 2021 Wenhan Xiong, Barlas Oğuz, Anchit Gupta, Xilun Chen, Diana Liskovich, Omer Levy, Wen-tau Yih, Yashar Mehdad

Many NLP tasks require processing long contexts beyond the length limit of pretrained models.

Salient Phrase Aware Dense Retrieval: Can a Dense Retriever Imitate a Sparse One?

1 code implementation13 Oct 2021 Xilun Chen, Kushal Lakhotia, Barlas Oğuz, Anchit Gupta, Patrick Lewis, Stan Peshterliev, Yashar Mehdad, Sonal Gupta, Wen-tau Yih

Despite their recent popularity and well-known advantages, dense retrievers still lag behind sparse methods such as BM25 in their ability to reliably match salient phrases and rare entities in the query and to generalize to out-of-domain data.

Open-Domain Question Answering Passage Retrieval

Domain-matched Pre-training Tasks for Dense Retrieval

1 code implementation28 Jul 2021 Barlas Oğuz, Kushal Lakhotia, Anchit Gupta, Patrick Lewis, Vladimir Karpukhin, Aleksandra Piktus, Xilun Chen, Sebastian Riedel, Wen-tau Yih, Sonal Gupta, Yashar Mehdad

Pre-training on larger datasets with ever increasing model size is now a proven recipe for increased performance across almost all NLP tasks.

 Ranked #1 on Passage Retrieval on Natural Questions (using extra training data)

Passage Retrieval

Low-Resource Domain Adaptation for Compositional Task-Oriented Semantic Parsing

no code implementations EMNLP 2020 Xilun Chen, Asish Ghoshal, Yashar Mehdad, Luke Zettlemoyer, Sonal Gupta

Task-oriented semantic parsing is a critical component of virtual assistants, which is responsible for understanding the user's intents (set reminder, play music, etc.).

Domain Adaptation Meta-Learning +2

Multi-Source Cross-Lingual Model Transfer: Learning What to Share

1 code implementation ACL 2019 Xilun Chen, Ahmed Hassan Awadallah, Hany Hassan, Wei Wang, Claire Cardie

In this work, we focus on the multilingual transfer setting where training data in multiple source languages is leveraged to further boost target language performance.

Cross-Lingual NER Text Classification +1

Zero-Resource Multilingual Model Transfer: Learning What to Share

no code implementations27 Sep 2018 Xilun Chen, Ahmed Hassan Awadallah, Hany Hassan, Wei Wang, Claire Cardie

In this work, we propose a zero-resource multilingual transfer learning model that can utilize training data in multiple source languages, while not requiring target language training data nor cross-lingual supervision.

Cross-Lingual Transfer Natural Language Processing +2

Unsupervised Multilingual Word Embeddings

3 code implementations EMNLP 2018 Xilun Chen, Claire Cardie

Multilingual Word Embeddings (MWEs) represent words from multiple languages in a single distributional vector space.

Multilingual Word Embeddings Translation +2

Adversarial Deep Averaging Networks for Cross-Lingual Sentiment Classification

2 code implementations TACL 2018 Xilun Chen, Yu Sun, Ben Athiwaratkun, Claire Cardie, Kilian Weinberger

To tackle the sentiment classification problem in low-resource languages without adequate annotated data, we propose an Adversarial Deep Averaging Network (ADAN) to transfer the knowledge learned from labeled data on a resource-rich source language to low-resource languages where only unlabeled data exists.

Classification Cross-Lingual Document Classification +4

Cannot find the paper you are looking for? You can Submit a new open access paper.