no code implementations • EMNLP (ALW) 2020 • Vinodkumar Prabhakaran, Zeerak Waseem, Seyi Akiwowo, Bertie Vidgen
In 2020 The Workshop on Online Abuse and Harms (WOAH) held a satellite panel at RightsCons 2020, an international human rights conference.
no code implementations • LREC 2022 • Jennifer Tracey, Owen Rambow, Claire Cardie, Adam Dalton, Hoa Trang Dang, Mona Diab, Bonnie Dorr, Louise Guthrie, Magdalena Markowska, Smaranda Muresan, Vinodkumar Prabhakaran, Samira Shaikh, Tomek Strzalkowski
We present the BeSt corpus, which records cognitive state: who believes what (i. e., factuality), and who has what sentiment towards what.
no code implementations • ACL (WOAH) 2021 • Lambert Mathias, Shaoliang Nie, Aida Mostafazadeh Davani, Douwe Kiela, Vinodkumar Prabhakaran, Bertie Vidgen, Zeerak Waseem
We present the results and main findings of the shared task at WOAH 5 on hateful memes detection.
no code implementations • 19 Feb 2025 • Piyawat Lertvittayakumjorn, David Kinney, Vinodkumar Prabhakaran, Donald Martin, Sunipa Dev
Specifically, we compare the performance of standard LLMs, LLMs augmented with retrievals from a bespoke knowledge base (i. e., KB grounding), and LLMs augmented with retrievals from a web search (i. e., search grounding) on a series of cultural familiarity benchmarks.
no code implementations • 2 Jan 2025 • Rida Qadri, Aida M. Davani, Kevin Robinson, Vinodkumar Prabhakaran
As a result, language models will shape how people learn about, perceive and interact with global cultures making it important to consider whose knowledge systems and perspectives are represented in models.
no code implementations • 22 Oct 2024 • Charvi Rastogi, Tian Huey Teh, Pushkar Mishra, Roma Patel, Zoe Ashwood, Aida Mostafazadeh Davani, Mark Diaz, Michela Paganini, Alicia Parrish, Ding Wang, Vinodkumar Prabhakaran, Lora Aroyo, Verena Rieser
Our study shows that (1) there are significant differences across demographic groups (including intersectional groups) on how severe they assess the harm to be, and that these differences vary across different types of safety violations, (2) the diverse rater pool captures annotation patterns that are substantially different from expert raters trained on specific set of safety policies, and (3) the differences we observe in T2I safety are distinct from previously documented group level differences in text-based safety tasks.
2 code implementations • 9 Jul 2024 • Nithish Kannen, Arif Ahmad, Marco Andreetto, Vinodkumar Prabhakaran, Utsav Prabhu, Adji Bousso Dieng, Pushpak Bhattacharyya, Shachi Dave
CUBE consists of 1) CUBE-1K, a set of high-quality prompts that enable the evaluation of cultural awareness, and 2) CUBE-CSpace, a larger dataset of cultural artifacts that serves as grounding to evaluate cultural diversity.
no code implementations • 16 Apr 2024 • Aida Mostafazadeh Davani, Mark Díaz, Dylan Baker, Vinodkumar Prabhakaran
While human annotations play a crucial role in language technologies, annotator subjectivity has long been overlooked in data collection.
no code implementations • 8 Apr 2024 • Aida Mostafazadeh Davani, Sagar Gubbi, Sunipa Dev, Shachi Dave, Vinodkumar Prabhakaran
We argue that understanding the sentential context is crucial for detecting instances of generalization.
1 code implementation • 8 Mar 2024 • Mukul Bhutani, Kevin Robinson, Vinodkumar Prabhakaran, Shachi Dave, Sunipa Dev
While generative multilingual models are rapidly being deployed, their safety and fairness evaluations are largely limited to resources collected in English.
1 code implementation • 12 Jan 2024 • Akshita Jha, Vinodkumar Prabhakaran, Remi Denton, Sarah Laszlo, Shachi Dave, Rida Qadri, Chandan K. Reddy, Sunipa Dev
First, we show that stereotypical attributes in ViSAGe are thrice as likely to be present in generated images of corresponding identities as compared to other attributes, and that the offensiveness of these depictions is especially higher for identities from Africa, South America, and South East Asia.
no code implementations • 11 Dec 2023 • Aida Davani, Mark Díaz, Dylan Baker, Vinodkumar Prabhakaran
More importantly, we find that individual moral values play a crucial role in shaping these variations: moral concerns about Care and Purity are significant mediating factors driving cross-cultural differences.
no code implementations • 28 Nov 2023 • Mark Díaz, Sunipa Dev, Emily Reif, Emily Denton, Vinodkumar Prabhakaran
The unstructured nature of data used in foundation model development is a challenge to systematic analyses for making data use and documentation decisions.
no code implementations • 9 Nov 2023 • Vinodkumar Prabhakaran, Christopher Homan, Lora Aroyo, Aida Mostafazadeh Davani, Alicia Parrish, Alex Taylor, Mark Díaz, Ding Wang, Gregory Serapio-García
Human annotation plays a core role in machine learning -- annotations for supervised models, safety guardrails for generative models, and human feedback for reinforcement learning, to cite a few avenues.
no code implementations • 7 Nov 2023 • Wenbo Zhang, Hangzhi Guo, Ian D Kivlichan, Vinodkumar Prabhakaran, Davis Yadav, Amulya Yadav
Toxicity is an increasingly common and severe issue in online spaces.
1 code implementation • 19 May 2023 • Akshita Jha, Aida Davani, Chandan K. Reddy, Shachi Dave, Vinodkumar Prabhakaran, Sunipa Dev
Stereotype benchmark datasets are crucial to detect and mitigate social stereotypes about groups of people in NLP models.
no code implementations • 19 May 2023 • Jacob Eisenstein, Vinodkumar Prabhakaran, Clara Rivera, Dorottya Demszky, Devyani Sharma
We introduce a new dataset of conversational speech representing English from India, Nigeria, and the United States.
no code implementations • 21 Nov 2022 • Shaily Bhatt, Sunipa Dev, Partha Talukdar, Shachi Dave, Vinodkumar Prabhakaran
Recent research has revealed undesirable biases in NLP data and models.
no code implementations • 19 Nov 2022 • Vinodkumar Prabhakaran, Rida Qadri, Ben Hutchinson
Artificial intelligence (AI) systems attempt to imitate human behavior.
no code implementations • 11 Oct 2022 • Ben Hutchinson, Jason Baldridge, Vinodkumar Prabhakaran
Questions regarding implicitness, ambiguity and underspecification are crucial for understanding the task validity and ethical concerns of multimodal image+text systems, yet have received little attention to date.
no code implementations • 6 Oct 2022 • Vinodkumar Prabhakaran, Margaret Mitchell, Timnit Gebru, Iason Gabriel
Research on fairness, accountability, transparency and ethics of AI-based interventions in society has gained much-needed momentum in recent years.
1 code implementation • 25 Sep 2022 • Shaily Bhatt, Sunipa Dev, Partha Talukdar, Shachi Dave, Vinodkumar Prabhakaran
In this paper, we focus on NLP fair-ness in the context of India.
no code implementations • 9 Jun 2022 • Mark Diaz, Ian D. Kivlichan, Rachel Rosen, Dylan K. Baker, Razvan Amironesei, Vinodkumar Prabhakaran, Emily Denton
Human annotated data plays a crucial role in machine learning (ML) research and development.
no code implementations • 11 May 2022 • Ben Hutchinson, Negar Rostamzadeh, Christina Greer, Katherine Heller, Vinodkumar Prabhakaran
Forming a reliable judgement of a machine learning (ML) model's appropriateness for an application ecosystem is critical for its responsible use, and requires considering a broad range of factors including harms, benefits, and responsibilities.
7 code implementations • Google Research 2022 • Aakanksha Chowdhery, Sharan Narang, Jacob Devlin, Maarten Bosma, Gaurav Mishra, Adam Roberts, Paul Barham, Hyung Won Chung, Charles Sutton, Sebastian Gehrmann, Parker Schuh, Kensen Shi, Sasha Tsvyashchenko, Joshua Maynez, Abhishek Rao, Parker Barnes, Yi Tay, Noam Shazeer, Vinodkumar Prabhakaran, Emily Reif, Nan Du, Ben Hutchinson, Reiner Pope, James Bradbury, Jacob Austin, Michael Isard, Guy Gur-Ari, Pengcheng Yin, Toju Duke, Anselm Levskaya, Sanjay Ghemawat, Sunipa Dev, Henryk Michalewski, Xavier Garcia, Vedant Misra, Kevin Robinson, Liam Fedus, Denny Zhou, Daphne Ippolito, David Luan, Hyeontaek Lim, Barret Zoph, Alexander Spiridonov, Ryan Sepassi, David Dohan, Shivani Agrawal, Mark Omernick, Andrew M. Dai, Thanumalayan Sankaranarayana Pillai, Marie Pellat, Aitor Lewkowycz, Erica Moreira, Rewon Child, Oleksandr Polozov, Katherine Lee, Zongwei Zhou, Xuezhi Wang, Brennan Saeta, Mark Diaz, Orhan Firat, Michele Catasta, Jason Wei, Kathy Meier-Hellstern, Douglas Eck, Jeff Dean, Slav Petrov, Noah Fiedel
To further our understanding of the impact of scale on few-shot learning, we trained a 540-billion parameter, densely activated, Transformer language model, which we call Pathways Language Model PaLM.
Ranked #1 on
Coreference Resolution
on Winograd Schema Challenge
2 code implementations • 20 Jan 2022 • Romal Thoppilan, Daniel De Freitas, Jamie Hall, Noam Shazeer, Apoorv Kulshreshtha, Heng-Tze Cheng, Alicia Jin, Taylor Bos, Leslie Baker, Yu Du, Yaguang Li, Hongrae Lee, Huaixiu Steven Zheng, Amin Ghafouri, Marcelo Menegali, Yanping Huang, Maxim Krikun, Dmitry Lepikhin, James Qin, Dehao Chen, Yuanzhong Xu, Zhifeng Chen, Adam Roberts, Maarten Bosma, Vincent Zhao, Yanqi Zhou, Chung-Ching Chang, Igor Krivokon, Will Rusch, Marc Pickett, Pranesh Srinivasan, Laichee Man, Kathleen Meier-Hellstern, Meredith Ringel Morris, Tulsee Doshi, Renelito Delos Santos, Toju Duke, Johnny Soraker, Ben Zevenbergen, Vinodkumar Prabhakaran, Mark Diaz, Ben Hutchinson, Kristen Olson, Alejandra Molina, Erin Hoffman-John, Josh Lee, Lora Aroyo, Ravi Rajakumar, Alena Butryna, Matthew Lamm, Viktoriya Kuzmina, Joe Fenton, Aaron Cohen, Rachel Bernstein, Ray Kurzweil, Blaise Aguera-Arcas, Claire Cui, Marian Croak, Ed Chi, Quoc Le
We demonstrate that fine-tuning with annotated data and enabling the model to consult external knowledge sources can lead to significant improvements towards the two key challenges of safety and factual grounding.
no code implementations • 8 Dec 2021 • Emily Denton, Mark Díaz, Ian Kivlichan, Vinodkumar Prabhakaran, Rachel Rosen
Human annotations play a crucial role in machine learning (ML) research and development.
no code implementations • 6 Dec 2021 • Negar Rostamzadeh, Ben Hutchinson, Christina Greer, Vinodkumar Prabhakaran
Testing practices within the machine learning (ML) community have centered around assessing a learned model's predictive performance measured against a test dataset, often drawn from the same distribution as the training dataset.
1 code implementation • 12 Oct 2021 • Aida Mostafazadeh Davani, Mark Díaz, Vinodkumar Prabhakaran
Majority voting and averaging are common approaches employed to resolve annotator disagreements and derive single ground truth labels from multiple annotations.
no code implementations • EMNLP (LAW, DMR) 2021 • Vinodkumar Prabhakaran, Aida Mostafazadeh Davani, Mark Díaz
A common practice in building NLP datasets, especially using crowd-sourced annotations, involves obtaining multiple annotator judgements on the same data instances, which are then flattened to produce a single "ground truth" label or score, through majority voting, averaging, or adjudication.
no code implementations • WNUT (ACL) 2021 • Sayan Ghosh, Dylan Baker, David Jurgens, Vinodkumar Prabhakaran
Online social media platforms increasingly rely on Natural Language Processing (NLP) techniques to detect abusive content at scale in order to mitigate the harms it causes to their users.
no code implementations • 8 Apr 2021 • Vinodkumar Prabhakaran, Marek Rei, Ekaterina Shutova
Metaphors are widely used in political rhetoric as an effective framing device.
no code implementations • 25 Jan 2021 • Nithya Sambasivan, Erin Arnesen, Ben Hutchinson, Tulsee Doshi, Vinodkumar Prabhakaran
Instead, we re-imagine algorithmic fairness in India and provide a roadmap to re-contextualise data and models, empower oppressed communities, and enable Fair-ML ecosystems.
no code implementations • 3 Dec 2020 • Nithya Sambasivan, Erin Arnesen, Ben Hutchinson, Vinodkumar Prabhakaran
Conventional algorithmic fairness is Western in its sub-groups, values, and optimizations.
no code implementations • NAACL 2021 • Dorottya Demszky, Devyani Sharma, Jonathan H. Clark, Vinodkumar Prabhakaran, Jacob Eisenstein
Evaluation on a test set of 22 dialect features of Indian English demonstrates that these models learn to recognize many features with high accuracy, and that a few minimal pairs can be as effective for training as thousands of labeled examples.
no code implementations • 17 Jun 2020 • Donald Martin Jr., Vinodkumar Prabhakaran, Jill Kuhlberg, Andrew Smart, William S. Isaac
Machine learning (ML) fairness research tends to focus primarily on mathematically-based interventions on often opaque algorithms or models and/or their immediate inputs and outputs.
no code implementations • 15 May 2020 • Donald Martin Jr., Vinodkumar Prabhakaran, Jill Kuhlberg, Andrew Smart, William S. Isaac
Recent research on algorithmic fairness has highlighted that the problem formulation phase of ML system development can be a key source of bias that has significant downstream impacts on ML system fairness outcomes.
no code implementations • ACL 2020 • Ben Hutchinson, Vinodkumar Prabhakaran, Emily Denton, Kellie Webster, Yu Zhong, Stephen Denuyl
Building equitable and inclusive NLP technologies demands consideration of whether and how social attitudes are represented in ML models.
no code implementations • IJCNLP 2019 • Vinodkumar Prabhakaran, Ben Hutchinson, Margaret Mitchell
Data-driven statistical Natural Language Processing (NLP) techniques leverage large amounts of language data to build models that can understand language.
no code implementations • COLING 2018 • Michelle Lam, Catherina Xu, Angela Kong, Vinodkumar Prabhakaran
Can language analysis reveal the underlying social power relations that exist between participants of an interaction?
no code implementations • NAACL 2018 • Yulia Tsvetkov, Vinodkumar Prabhakaran, Rob Voigt
As language technologies have become increasingly prevalent, there is a growing awareness that decisions we make about our data, methods, and tools are often tied up with their impact on people and societies.
no code implementations • NAACL 2018 • Vinodkumar Prabhakaran, Premkumar Ganeshkumar, Owen Rambow
Understanding how social power structures affect the way we interact with one another is of great interest to social scientists who want to answer fundamental questions about human behavior, as well as to computer scientists who want to build automatic methods to infer the social contexts of interactions.
no code implementations • TACL 2018 • Vinodkumar Prabhakaran, Camilla Griffiths, Hang Su, Prateek Verma, Nelson Morgan, Jennifer L. Eberhardt, Dan Jurafsky
We apply computational dialog methods to police body-worn camera footage to model conversations between police officers and community members in traffic stops.
no code implementations • 12 Jun 2017 • Vinodkumar Prabhakaran, Owen Rambow
In this paper, we study the interaction of power, gender, and dialog behavior in organizational interactions.
no code implementations • EACL 2017 • Henning Wachsmuth, Nona Naderi, Yufang Hou, Yonatan Bilu, Vinodkumar Prabhakaran, Tim Alberdingk Thijm, Graeme Hirst, Benno Stein
Research on computational argumentation faces the problem of how to automatically assess the quality of an argument or argumentation.
no code implementations • LREC 2016 • Vinodkumar Prabhakaran, Owen Rambow
In order to gain a deep understanding of how social context manifests in interactions, we need data that represents interactions from a large community of people over a long period of time, capturing different aspects of social context.
no code implementations • SEMEVAL 2015 • Vinodkumar Prabhakaran, Tomas By, Julia Hirschberg, Owen Rambow, Samira Shaikh, Tomek Strzalkowski, Jennifer Tracey, Michael Arrigo, Rupayan Basu, Micah Clark, Adam Dalton, Mona Diab, Louise Guthrie, Anna Prokofieva, Stephanie Strassel, Gregory Werner, Yorick Wilks, Janyce Wiebe
no code implementations • WS 2012 • Vinodkumar Prabhakaran, Michael Bloodgood, Mona Diab, Bonnie Dorr, Lori Levin, Christine D. Piatko, Owen Rambow, Benjamin Van Durme
We explore training an automatic modality tagger.
no code implementations • LREC 2012 • Vinodkumar Prabhakaran, Huzaifa Neralwala, Owen Rambow, Mona Diab
In this paper, we describe a multi-layer annotation scheme for social power relations that are recognizable from online written interactions.