no code implementations • NLPerspectives (LREC) 2022 • Christopher Homan, Tharindu Cyril Weerasooriya, Lora Aroyo, Chris Welty
Annotator disagreement is often dismissed as noise or the result of poor annotation process quality.
1 code implementation • 22 Aug 2023 • Oana Inel, Tim Draws, Lora Aroyo
We argue that data collection for AI should be performed in a responsible manner where the quality of the data is thoroughly scrutinized and measured through a systematic set of appropriate metrics.
no code implementations • 27 Jun 2023 • Alicia Parrish, Sarah Laszlo, Lora Aroyo
Many questions that we ask about the world do not have a single clear answer, yet typical human annotation set-ups in machine learning assume there must be a single ground truth label for all examples in every task.
no code implementations • 22 May 2023 • Alicia Parrish, Hannah Rose Kirk, Jessica Quaye, Charvi Rastogi, Max Bartolo, Oana Inel, Juan Ciro, Rafael Mosquera, Addison Howard, Will Cukierski, D. Sculley, Vijay Janapa Reddi, Lora Aroyo
To address this need, we introduce the Adversarial Nibbler challenge.
no code implementations • 16 Feb 2023 • Mohammad Tahaei, Marios Constantinides, Daniele Quercia, Sean Kennedy, Michael Muller, Simone Stumpf, Q. Vera Liao, Ricardo Baeza-Yates, Lora Aroyo, Jess Holbrook, Ewa Luger, Michael Madaio, Ilana Golbin Blumenfeld, Maria De-Arteaga, Jessica Vitak, Alexandra Olteanu
In recent years, the CHI community has seen significant growth in research on Human-Centered Responsible Artificial Intelligence.
1 code implementation • 20 Jul 2022 • Mark Mazumder, Colby Banbury, Xiaozhe Yao, Bojan Karlaš, William Gaviria Rojas, Sudnya Diamos, Greg Diamos, Lynn He, Alicia Parrish, Hannah Rose Kirk, Jessica Quaye, Charvi Rastogi, Douwe Kiela, David Jurado, David Kanter, Rafael Mosquera, Juan Ciro, Lora Aroyo, Bilge Acun, Lingjiao Chen, Mehul Smriti Raje, Max Bartolo, Sabri Eyuboglu, Amirata Ghorbani, Emmett Goodman, Oana Inel, Tariq Kane, Christine R. Kirkpatrick, Tzu-Sheng Kuo, Jonas Mueller, Tristan Thrush, Joaquin Vanschoren, Margaret Warren, Adina Williams, Serena Yeung, Newsha Ardalani, Praveen Paritosh, Ce Zhang, James Zou, Carole-Jean Wu, Cody Coleman, Andrew Ng, Peter Mattson, Vijay Janapa Reddi
Machine learning research has long focused on models rather than datasets, and prominent datasets are used for common ML tasks without regard to the breadth, difficulty, and faithfulness of the underlying problems.
2 code implementations • 20 Jan 2022 • Romal Thoppilan, Daniel De Freitas, Jamie Hall, Noam Shazeer, Apoorv Kulshreshtha, Heng-Tze Cheng, Alicia Jin, Taylor Bos, Leslie Baker, Yu Du, Yaguang Li, Hongrae Lee, Huaixiu Steven Zheng, Amin Ghafouri, Marcelo Menegali, Yanping Huang, Maxim Krikun, Dmitry Lepikhin, James Qin, Dehao Chen, Yuanzhong Xu, Zhifeng Chen, Adam Roberts, Maarten Bosma, Vincent Zhao, Yanqi Zhou, Chung-Ching Chang, Igor Krivokon, Will Rusch, Marc Pickett, Pranesh Srinivasan, Laichee Man, Kathleen Meier-Hellstern, Meredith Ringel Morris, Tulsee Doshi, Renelito Delos Santos, Toju Duke, Johnny Soraker, Ben Zevenbergen, Vinodkumar Prabhakaran, Mark Diaz, Ben Hutchinson, Kristen Olson, Alejandra Molina, Erin Hoffman-John, Josh Lee, Lora Aroyo, Ravi Rajakumar, Alena Butryna, Matthew Lamm, Viktoriya Kuzmina, Joe Fenton, Aaron Cohen, Rachel Bernstein, Ray Kurzweil, Blaise Aguera-Arcas, Claire Cui, Marian Croak, Ed Chi, Quoc Le
We demonstrate that fine-tuning with annotated data and enabling the model to consult external knowledge sources can lead to significant improvements towards the two key challenges of safety and factual grounding.
Ranked #40 on
Code Generation
on HumanEval
1 code implementation • 23 Dec 2021 • Hannah Rashkin, Vitaly Nikolaev, Matthew Lamm, Lora Aroyo, Michael Collins, Dipanjan Das, Slav Petrov, Gaurav Singh Tomar, Iulia Turc, David Reitter
With recent improvements in natural language generation (NLG) models for various applications, it has become imperative to have the means to identify and evaluate whether NLG output is only sharing verifiable information about the external world.
no code implementations • 19 Nov 2021 • Lora Aroyo, Matthew Lease, Praveen Paritosh, Mike Schaekermann
The efficacy of machine learning (ML) models depends on both algorithms and data.
no code implementations • ACL 2021 • Ka Wong, Praveen Paritosh, Lora Aroyo
When collecting annotations and labeled data from humans, a standard practice is to use inter-rater reliability (IRR) as a measure of data goodness (Hallgren, 2012).
no code implementations • 11 Jun 2021 • Ka Wong, Praveen Paritosh, Lora Aroyo
We present a new approach to interpreting IRR that is empirical and contextualized.
1 code implementation • 5 Nov 2019 • Chris Welty, Praveen Paritosh, Lora Aroyo
In this paper we present the first steps towards hardening the science of measuring AI systems, by adopting metrology, the science of measurement and its application, and applying it to human (crowd) powered evaluations.
1 code implementation • NAACL 2019 • Anca Dumitrache, Lora Aroyo, Chris Welty
We present a resource for the task of FrameNet semantic frame disambiguation of over 5, 000 word-sentence pairs from the Wikipedia corpus.
no code implementations • WS 2018 • Sven van den Beukel, Lora Aroyo
In this paper, automatic homophone- and homograph detection are suggested as new useful features for humor recognition systems.
1 code implementation • WS 2018 • Anca Dumitrache, Lora Aroyo, Chris Welty
Distant supervision is a popular method for performing relation extraction from text that is known to produce noisy labels.
2 code implementations • 18 Aug 2018 • Anca Dumitrache, Oana Inel, Lora Aroyo, Benjamin Timmermans, Chris Welty
However, in many domains, there is ambiguity in the data, as well as a multitude of perspectives of the information examples.
Human-Computer Interaction Social and Information Networks
1 code implementation • COLING 2018 • Chantal van Son, Roser Morante, Lora Aroyo, Piek Vossen
This paper reports on a reimplementation of a system on detecting implicit positive meaning from negated statements.
1 code implementation • 1 May 2018 • Anca Dumitrache, Lora Aroyo, Chris Welty
FrameNet is a computational linguistics resource composed of semantic frames, high-level concepts that represent the meanings of words.
1 code implementation • 14 Nov 2017 • Anca Dumitrache, Lora Aroyo, Chris Welty
Distant supervision (DS) is a well-established method for relation extraction from text, based on the assumption that when a knowledge-base contains a relation between a term pair, then sentences that contain that pair are likely to express the relation.
1 code implementation • 9 Jan 2017 • Anca Dumitrache, Lora Aroyo, Chris Welty
Cognitive computing systems require human labeled data for evaluation, and often for training.
no code implementations • LREC 2016 • Oana Inel, Tommaso Caselli, Lora Aroyo
On the other hand, machines need to understand the information that is published in online data streams and generate concise and meaningful overviews.
no code implementations • LREC 2016 • Emiel van Miltenburg, Benjamin Timmermans, Lora Aroyo
The main goal of this study is to find out (i) whether it is feasible to collect keywords for a large collection of sounds through crowdsourcing, and (ii) how people talk about sounds, and what information they can infer from hearing a sound in isolation.
no code implementations • LREC 2016 • Chantal van Son, Tommaso Caselli, Antske Fokkens, Isa Maks, Roser Morante, Lora Aroyo, Piek Vossen
In the last decade, different aspects of linguistic encoding of perspectives have been targeted as separated phenomena through different annotation initiatives.