1 code implementation • 16 Feb 2023 • Tomasz Korbak, Kejian Shi, Angelica Chen, Rasika Bhalerao, Christopher L. Buckley, Jason Phang, Samuel R. Bowman, Ethan Perez
Language models (LMs) are pretrained to imitate internet text, including content that would violate human preferences if generated by an LM: falsehoods, offensive comments, personally identifiable information, low-quality or buggy code, and more.
no code implementations • 19 Feb 2022 • Rasika Bhalerao, Mohammad Al-Rubaie, Anand Bhaskar, Igor Markov
We propose Continuous Word2Vec (CW2V), our data-driven method to learn word embeddings that ensures that perturbations of words have embeddings similar to those of the original words.
1 code implementation • EMNLP 2020 • Nikita Nangia, Clara Vania, Rasika Bhalerao, Samuel R. Bowman
To measure some forms of social bias in language models against protected demographic groups in the US, we introduce the Crowdsourced Stereotype Pairs benchmark (CrowS-Pairs).
no code implementations • 2 Dec 2018 • Rasika Bhalerao, Maxwell Aliapoulios, Ilia Shumailov, Sadia Afroz, Damon McCoy
Our analysis of the automatically generated supply chains demonstrates underlying connections between products and services within these forums.