Search Results for author: Lingfeng Shen

Found 14 papers, 6 papers with code

AnaloBench: Benchmarking the Identification of Abstract and Long-context Analogies

1 code implementation19 Feb 2024 Xiao Ye, Andrew Wang, Jacob Choi, Yining Lu, Shreya Sharma, Lingfeng Shen, Vijay Tiyyala, Nicholas Andrews, Daniel Khashabi

Our benchmarking approach focuses on aspects of this ability that are common among humans: (i) recalling related experiences from a large amount of information, and (ii) applying analogical reasoning to complex and lengthy scenarios.

Benchmarking

The Language Barrier: Dissecting Safety Challenges of LLMs in Multilingual Contexts

no code implementations23 Jan 2024 Lingfeng Shen, Weiting Tan, Sihao Chen, Yunmo Chen, Jingyu Zhang, Haoran Xu, Boyuan Zheng, Philipp Koehn, Daniel Khashabi

As the influence of large language models (LLMs) spans across global communities, their safety challenges in multilingual settings become paramount for alignment research.

Contrastive Preference Optimization: Pushing the Boundaries of LLM Performance in Machine Translation

1 code implementation16 Jan 2024 Haoran Xu, Amr Sharaf, Yunmo Chen, Weiting Tan, Lingfeng Shen, Benjamin Van Durme, Kenton Murray, Young Jin Kim

However, even the top-performing 13B LLM-based translation models, like ALMA, does not match the performance of state-of-the-art conventional encoder-decoder translation models or larger-scale LLMs such as GPT-4.

Machine Translation Translation

Narrowing the Gap between Zero- and Few-shot Machine Translation by Matching Styles

no code implementations4 Nov 2023 Weiting Tan, Haoran Xu, Lingfeng Shen, Shuyue Stella Li, Kenton Murray, Philipp Koehn, Benjamin Van Durme, Yunmo Chen

Large language models trained primarily in a monolingual setting have demonstrated their ability to generalize to machine translation using zero- and few-shot examples with in-context learning.

In-Context Learning Machine Translation +1

Revisiting the Hypothesis: Do pretrained Transformers Learn In-Context by Gradient Descent?

no code implementations12 Oct 2023 Lingfeng Shen, Aayush Mishra, Daniel Khashabi

We observe that ICL and GD have different sensitivity to the order in which they observe demonstrations.

In-Context Learning

Sen2Pro: A Probabilistic Perspective to Sentence Embedding from Pre-trained Language Model

no code implementations4 Jun 2023 Lingfeng Shen, Haiyun Jiang, Lemao Liu, Shuming Shi

Sentence embedding is one of the most fundamental tasks in Natural Language Processing and plays an important role in various tasks.

Language Modelling Sentence +2

Flatness-Aware Prompt Selection Improves Accuracy and Sample Efficiency

1 code implementation18 May 2023 Lingfeng Shen, Weiting Tan, Boyuan Zheng, Daniel Khashabi

We provide theoretical foundations for this metric and its relationship with other prompt selection metrics, providing a comprehensive understanding of existing methods.

A Simple and Plug-and-play Method for Unsupervised Sentence Representation Enhancement

no code implementations13 May 2023 Lingfeng Shen, Haiyun Jiang, Lemao Liu, Shuming Shi

Generating proper embedding of sentences through an unsupervised way is beneficial to semantic matching and retrieval problems in real-world scenarios.

Retrieval Sentence +2

Frequency-aware Dimension Selection for Static Word Embedding by Mixed Product Distance

no code implementations13 May 2023 Lingfeng Shen, Haiyun Jiang, Lemao Liu, Ying Chen

Static word embedding is still useful, particularly for context-unavailable tasks, because in the case of no context available, pre-trained language models often perform worse than static word embeddings.

Word Embeddings

On the Evaluation Metrics for Paraphrase Generation

1 code implementation17 Feb 2022 Lingfeng Shen, Lemao Liu, Haiyun Jiang, Shuming Shi

In this paper we revisit automatic metrics for paraphrase evaluation and obtain two findings that disobey conventional wisdom: (1) Reference-free metrics achieve better performance than their reference-based counterparts.

Machine Translation Paraphrase Generation

Rethink the Evaluation for Attack Strength of Backdoor Attacks in Natural Language Processing

no code implementations9 Jan 2022 Lingfeng Shen, Haiyun Jiang, Lemao Liu, Shuming Shi

It has been shown that natural language processing (NLP) models are vulnerable to a kind of security threat called the Backdoor Attack, which utilizes a `backdoor trigger' paradigm to mislead the models.

Backdoor Attack Text Classification

Cannot find the paper you are looking for? You can Submit a new open access paper.