no code implementations • EMNLP (BlackboxNLP) 2021 • Zhiying Jiang, Raphael Tang, Ji Xin, Jimmy Lin
Fine-tuned pre-trained transformers achieve the state of the art in passage reranking.
no code implementations • ACL (NLP4Prog) 2021 • Xinyu Zhang, Ji Xin, Andrew Yates, Jimmy Lin
The task of semantic code search is to retrieve code snippets from a source code corpus based on an information need expressed in natural language.
no code implementations • EMNLP 2021 • Raphael Tang, Karun Kumar, Kendra Chalkley, Ji Xin, Liming Zhang, Wenyan Li, Gefei Yang, Yajie Mao, Junho Shin, Geoffrey Craig Murray, Jimmy Lin
Query auto completion (QAC) is the task of predicting a search engine user’s final query from their intermediate, incomplete query.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
1 code implementation • EMNLP (sustainlp) 2020 • Ji Xin, Rodrigo Nogueira, YaoLiang Yu, Jimmy Lin
Pre-trained language models such as BERT have shown their effectiveness in various tasks.
no code implementations • EMNLP 2021 • Xueguang Ma, Minghan Li, Kai Sun, Ji Xin, Jimmy Lin
Recent work has shown that dense passage retrieval techniques achieve better ranking accuracy in open-domain question answering compared to sparse retrieval techniques such as BM25, but at the cost of large space and memory requirements.
no code implementations • 30 Nov 2023 • Linfeng Du, Ji Xin, Alex Labach, Saba Zuberi, Maksims Volkovs, Rahul G. Krishnan
Transformer-based models have greatly pushed the boundaries of time series forecasting recently.
no code implementations • 31 Jul 2022 • Ji Xin, Raphael Tang, Zhiying Jiang, YaoLiang Yu, Jimmy Lin
There exists a wide variety of efficiency methods for natural language processing (NLP) tasks, such as pruning, distillation, dynamic inference, quantization, etc.
no code implementations • 23 Jun 2022 • Zhiying Jiang, Yiqin Dai, Ji Xin, Ming Li, Jimmy Lin
Most real-world problems that machine learning algorithms are expected to solve face the situation with 1) unknown data distribution; 2) little domain-specific knowledge; and 3) datasets with limited annotation.
1 code implementation • 19 May 2022 • Minghan Li, Xinyu Zhang, Ji Xin, Hongyang Zhang, Jimmy Lin
For example, on MS MARCO Passage v1, our method yields an average candidate set size of 27 out of 1, 000 which increases the reranking speed by about 37 times, while the MRR@10 is greater than a pre-specified value of 0. 38 with about 90% empirical coverage and the empirical baselines fail to provide such guarantee.
no code implementations • Findings (ACL) 2022 • Ji Xin, Chenyan Xiong, Ashwin Srinivasan, Ankita Sharma, Damien Jose, Paul N. Bennett
Dense retrieval (DR) methods conduct text retrieval by first encoding texts in the embedding space and then matching them by nearest neighbor search.
no code implementations • 29 Sep 2021 • Ji Xin, Chenyan Xiong, Ashwin Srinivasan, Ankita Sharma, Damien Jose, Paul N. Bennett
Dense retrieval (DR) methods conduct text retrieval by first encoding texts in the embedding space and then matching them by nearest neighbor search.
1 code implementation • ACL 2021 • Ji Xin, Raphael Tang, YaoLiang Yu, Jimmy Lin
To fill this void in the literature, we study in this paper selective prediction for NLP, comparing different models and confidence estimators.
1 code implementation • EACL 2021 • Ji Xin, Raphael Tang, YaoLiang Yu, Jimmy Lin
The slow speed of BERT has motivated much research on accelerating its inference, and the early exiting idea has been proposed to make trade-offs between model quality and efficiency.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Zhiying Jiang, Raphael Tang, Ji Xin, Jimmy Lin
We show the effectiveness of our method in terms of attribution and the ability to provide insight into how information flows through layers.
1 code implementation • ACL 2020 • Raphael Tang, Jaejun Lee, Ji Xin, Xinyu Liu, Yao-Liang Yu, Jimmy Lin
In natural language processing, a recently popular line of work explores how to best report the experimental results of neural networks.
3 code implementations • ACL 2020 • Ji Xin, Raphael Tang, Jaejun Lee, Yao-Liang Yu, Jimmy Lin
Large-scale pre-trained language models such as BERT have brought significant improvements to NLP applications.
no code implementations • 15 Nov 2019 • Achyudh Ram, Ji Xin, Meiyappan Nagappan, Yao-Liang Yu, Rocío Cabrera Lozoya, Antonino Sabetta, Jimmy Lin
Public vulnerability databases such as CVE and NVD account for only 60% of security vulnerabilities present in open-source projects, and are known to suffer from inconsistent quality.
no code implementations • IJCNLP 2019 • Ji Xin, Jimmy Lin, Yao-Liang Yu
Memory neurons of long short-term memory (LSTM) networks encode and process information in powerful yet mysterious ways.
1 code implementation • EMNLP 2018 • Ji Xin, Hao Zhu, Xu Han, Zhiyuan Liu, Maosong Sun
Entity typing aims to classify semantic types of an entity mention in a specific context.