no code implementations • 19 Nov 2022 • Md Sultan Al Nahian, Spencer Frazier, Brent Harrison, Mark Riedl
To do this, we extend a dataset that has been previously used to train a binary normative classifier with annotations of moral principles.
no code implementations • 14 Oct 2022 • Louis Castricato, Alexander Havrilla, Shahbuland Matiana, Michael Pieler, Anbang Ye, Ian Yang, Spencer Frazier, Mark Riedl
However, simply fine-tuning a generative language model with a contrastive reward model does not always reliably result in a story generation system capable of generating stories that meet user preferences.
no code implementations • 7 Dec 2021 • Louis Castricato, Spencer Frazier, Jonathan Balloch, Nitya Tarakad, Mark Riedl
Neural language model-based approaches to automated story generation suffer from two important limitations.
no code implementations • 6 Oct 2021 • Shahbuland Matiana, JR Smith, Ryan Teehan, Louis Castricato, Stella Biderman, Leo Gao, Spencer Frazier
Recent advances in large-scale language models (Raffel et al., 2019; Brown et al., 2020) have brought significant qualitative and quantitative improvements in machine-driven text generation.
no code implementations • 19 Apr 2021 • Md Sultan Al Nahian, Spencer Frazier, Brent Harrison, Mark Riedl
As more machine learning agents interact with humans, it is increasingly a prospect that an agent trained to perform a task optimally, using only a measure of task performance as feedback, can violate societal norms for acceptable behavior or cause harm.
no code implementations • NAACL (NUSE) 2021 • Louis Castricato, Spencer Frazier, Jonathan Balloch, Mark Riedl
Automated story generation remains a difficult area of research because it lacks strong objective measures.
no code implementations • 4 Dec 2020 • Sahith Dambekodi, Spencer Frazier, Prithviraj Ammanabrolu, Mark O. Riedl
We test our technique in the 9to05 game, which is an extreme version of a text based game that requires numerous interactions with common, everyday objects in common, everyday scenarios.
no code implementations • INLG (ACL) 2020 • Xiangyu Peng, Siyan Li, Spencer Frazier, Mark Riedl
Our normative fine-tuning technique is able to reduce non-normative text by 27-61%, depending on the data set.
no code implementations • 7 Dec 2019 • Spencer Frazier, Md Sultan Al Nahian, Mark Riedl, Brent Harrison
Value alignment is a property of an intelligent agent indicating that it can only pursue goals and activities that are beneficial to humans.
no code implementations • 2 Aug 2019 • Spencer Frazier, Mark Riedl
We hypothesize that interactive machine learning IML, wherein human teachers play a direct role in training through demonstrations, critique, or action advice, may alleviate agent susceptibility to aliasing.