1 code implementation • 22 Jul 2024 • David Heineman, Yao Dou, Wei Xu
While instruction fine-tuned LLMs are effective text generators, sensitivity to prompt construction makes performance unstable and sub-optimal in practice.
no code implementations • 21 May 2024 • Govind Ramesh, Yao Dou, Wei Xu
Research on jailbreaking has been valuable for testing and understanding the safety and security issues of large language models (LLMs).
no code implementations • 16 Nov 2023 • Yao Dou, Isadora Krsek, Tarek Naous, Anubha Kabra, Sauvik Das, Alan Ritter, Wei Xu
Self-disclosure, while being common and rewarding in social media interaction, also poses privacy risks.
no code implementations • 5 Oct 2023 • Yao Dou, Philippe Laban, Claire Gardent, Wei Xu
In this tutorial, we focus on text-to-text generation, a class of natural language generation (NLG) tasks, that takes a piece of text as input and then generates a revision that is improved according to some specific criteria (e. g., readability or linguistic styles), while largely retaining the original meaning and the length of the text.
1 code implementation • 14 Aug 2023 • David Heineman, Yao Dou, Wei Xu
Additionally, we introduce a Python library to streamline the entire process from typology design and deployment to annotation processing.
no code implementations • 23 May 2023 • David Heineman, Yao Dou, Mounica Maddela, Wei Xu
Large language models (e. g., GPT-4) are uniquely capable of producing highly rated text simplification, yet current human evaluation methods fail to provide a clear understanding of systems' specific strengths and weaknesses.
1 code implementation • 19 Dec 2022 • Mounica Maddela, Yao Dou, David Heineman, Wei Xu
Training learnable metrics using modern language models has recently emerged as a promising method for the automatic evaluation of machine translation.
no code implementations • 6 Oct 2022 • Yao Dou, Chao Jiang, Wei Xu
This paper addresses the quality issues in existing Twitter-based paraphrase datasets, and discusses the necessity of using two separate definitions of paraphrase for identification and generation tasks.
no code implementations • ACL 2022 • Yao Dou, Maxwell Forbes, Rik Koncel-Kedziorski, Noah A. Smith, Yejin Choi
To support the broad range of real machine errors that can be identified by laypeople, the ten error categories of Scarecrow -- such as redundancy, commonsense errors, and incoherence -- are identified through several rounds of crowd annotation experiments without a predefined ontology.
no code implementations • 2 Feb 2021 • Yao Dou, Maxwell Forbes, Ari Holtzman, Yejin Choi
We study conversational dialog in which there are many possible responses to a given history.