Search Results for author: Michael Pieler

Found 9 papers, 5 papers with code

Arabic Stable LM: Adapting Stable LM 2 1.6B to Arabic

no code implementations5 Dec 2024 Zaid Alyafeai, Michael Pieler, Hannah Teufel, Jonathan Tow, Marco Bellagente, Duy Phung, Nikhil Pinnaparaju, Reshinth Adithyan, Paulo Rocha, Maksym Zhuravinskyi, Carlos Riquelme

Our Arabic Stable LM 1. 6B chat model achieves impressive results on several benchmarks beating multiple models with up to 8x the parameters.

Rephrasing natural text data with different languages and quality levels for Large Language Model pre-training

no code implementations28 Oct 2024 Michael Pieler, Marco Bellagente, Hannah Teufel, Duy Phung, Nathan Cooper, Jonathan Tow, Paulo Rocha, Reshinth Adithyan, Zaid Alyafeai, Nikhil Pinnaparaju, Maksym Zhuravinskyi, Carlos Riquelme

In addition, we provide a detailed study of our pipeline, investigating the choice of the base dataset and LLM for the rephrasing, as well as the relationship between the model size and the performance after pre-training.

Benchmarking Language Modelling +1

Robust Preference Learning for Storytelling via Contrastive Reinforcement Learning

no code implementations14 Oct 2022 Louis Castricato, Alexander Havrilla, Shahbuland Matiana, Michael Pieler, Anbang Ye, Ian Yang, Spencer Frazier, Mark Riedl

However, simply fine-tuning a generative language model with a contrastive reward model does not always reliably result in a story generation system capable of generating stories that meet user preferences.

Contrastive Learning Language Modelling +5

Few-shot Adaptation Works with UnpredicTable Data

1 code implementation1 Aug 2022 Jun Shern Chan, Michael Pieler, Jonathan Jao, Jérémy Scheurer, Ethan Perez

Finetuning on the resulting dataset leads to improved FSL performance on Natural Language Processing (NLP) tasks, but not proportionally to dataset scale.

Domain Adaptation Few-Shot Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.