no code implementations • 12 Sep 2023 • Angela Schöpke-Gonzalez, Siqi Wu, Sagar Kumar, Paul J. Resnick, Libby Hemphill
In designing instructions for annotation tasks to generate training data for these algorithms, researchers often treat the harm concepts that we train algorithms to detect - 'hateful', 'offensive', 'toxic', 'racist', 'sexist', etc.
1 code implementation • 24 May 2023 • MinJe Choi, Jiaxin Pei, Sagar Kumar, Chang Shu, David Jurgens
Large language models (LLMs) have been shown to perform well at a variety of syntactic, discourse, and reasoning tasks.