1 code implementation • 6 Mar 2024 • Luiza Pozzobon, Patrick Lewis, Sara Hooker, Beyza Ermis
To date, toxicity mitigation in language models has almost entirely been focused on single-language settings.
1 code implementation • 11 Oct 2023 • Luiza Pozzobon, Beyza Ermis, Patrick Lewis, Sara Hooker
Considerable effort has been dedicated to mitigating toxicity, but existing methods often require drastic modifications to model parameters or the use of computationally intensive auxiliary models.
no code implementations • 8 Sep 2023 • Max Marion, Ahmet Üstün, Luiza Pozzobon, Alex Wang, Marzieh Fadaee, Sara Hooker
In this work, we take a wider view and explore scalable estimates of data quality that can be used to systematically measure the quality of pretraining data.
1 code implementation • 24 Apr 2023 • Luiza Pozzobon, Beyza Ermis, Patrick Lewis, Sara Hooker
We evaluate the implications of these changes on the reproducibility of findings that compare the relative merits of models and methods that aim to curb toxicity.