no code implementations • EMNLP 2021 • Clara Meister, Afra Amini, Tim Vieira, Ryan Cotterell
Beam search is the default decoding strategy for many sequence generation tasks in NLP.
no code implementations • 8 Jul 2024 • Afra Amini, Tim Vieira, Ryan Cotterell
To the extent this fine-tuning is successful and we end up with a good approximation, we have reduced the inference cost by a factor of N. Our experiments on a controlled generation task suggest that while variational BoN is not as effective as BoN in aligning language models, it is close to BoN performance as vBoN appears more often on the Pareto frontier of reward and KL divergence compared to models trained with KL-constrained RL objective.
no code implementations • 25 Mar 2024 • Luca Malagutti, Andrius Buinovskij, Anej Svete, Clara Meister, Afra Amini, Ryan Cotterell
For nearly three decades, language models derived from the $n$-gram assumption held the state of the art on the task.
1 code implementation • 16 Feb 2024 • Afra Amini, Tim Vieira, Ryan Cotterell
DPO, as originally formulated, relies on binary preference data and fine-tunes a language model to increase the likelihood of a preferred response over a dispreferred response.
no code implementations • 29 Dec 2023 • Li Du, Afra Amini, Lucas Torroba Hennigen, Xinyan Velocity Yu, Jason Eisner, Holden Lee, Ryan Cotterell
Recent papers have demonstrated the possibility of energy-based text generation by adapting gradient-based sampling algorithms, a paradigm of MCMC algorithms that promises fast convergence.
no code implementations • 4 Oct 2023 • Jannis Bulian, Mike S. Schäfer, Afra Amini, Heidi Lam, Massimiliano Ciaramita, Ben Gaiarin, Michelle Chen Hübscher, Christian Buck, Niels G. Mede, Markus Leippold, Nadine Strauß
As Large Language Models (LLMs) rise in popularity, it is necessary to assess their capability in critically relevant domains.
no code implementations • 21 Jun 2023 • Robin Chan, Afra Amini, Mennatallah El-Assady
We present a human-in-the-loop dashboard tailored to diagnosing potential spurious features that NLI models rely on for predictions.
1 code implementation • 8 Jun 2023 • Afra Amini, Tianyu Liu, Ryan Cotterell
We introduce a novel dependency parser, the hexatagger, that constructs dependency trees by tagging the words in a sentence with elements from a finite set of possible tags.
1 code implementation • NeurIPS 2023 • Afra Amini, Li Du, Ryan Cotterell
In this paper, we take an important step toward building a principled approach for sampling from language models with gradient-based methods.
no code implementations • 24 May 2023 • Tianyu Liu, Afra Amini, Mrinmaya Sachan, Ryan Cotterell
We show that these exhaustive comparisons can be avoided, and, moreover, the complexity of such tasks can be reduced to linear by casting the relation between tokens as a partial order over the string.
no code implementations • 23 May 2023 • Afra Amini, Massimiliano Ciaramita
However, the effectiveness of in-context learning is dependent on the provided context, and the performance on a downstream task can vary considerably, depending on the instruction.
1 code implementation • 14 Nov 2022 • Afra Amini, Ryan Cotterell
There have been many proposals to reduce constituency parsing to tagging in the literature.
1 code implementation • 14 May 2022 • Afra Amini, Tiago Pimentel, Clara Meister, Ryan Cotterell
Probing has become a go-to methodology for interpreting and analyzing deep neural models in natural language processing.
1 code implementation • 22 Sep 2021 • Clara Meister, Afra Amini, Tim Vieira, Ryan Cotterell
In this work, we propose a new method for turning beam search into a stochastic process: Conditional Poisson stochastic beam search.