no code implementations • 22 Oct 2024 • Charvi Rastogi, Tian Huey Teh, Pushkar Mishra, Roma Patel, Zoe Ashwood, Aida Mostafazadeh Davani, Mark Diaz, Michela Paganini, Alicia Parrish, Ding Wang, Vinodkumar Prabhakaran, Lora Aroyo, Verena Rieser
Our study shows that (1) there are significant differences across demographic groups (including intersectional groups) on how severe they assess the harm to be, and that these differences vary across different types of safety violations, (2) the diverse rater pool captures annotation patterns that are substantially different from expert raters trained on specific set of safety policies, and (3) the differences we observe in T2I safety are distinct from previously documented group level differences in text-based safety tasks.
1 code implementation • 12 Oct 2024 • Ivo Verhoeven, Pushkar Mishra, Ekaterina Shutova
This paper introduces misinfo-general, a benchmark dataset for evaluating misinformation models' ability to perform out-of-distribution generalisation.
1 code implementation • 2 Apr 2024 • Ivo Verhoeven, Pushkar Mishra, Rahel Beloch, Helen Yannakoudakis, Ekaterina Shutova
This mismatch can be partially attributed to the limitations of current evaluation setups that neglect the rapid evolution of online content and the underlying social graph.
no code implementations • 24 Jul 2023 • Filippo Betello, Federico Siciliano, Pushkar Mishra, Fabrizio Silvestri
However, their robustness in the face of perturbations in training data remains a largely understudied yet critical issue.
19 code implementations • 18 Jul 2023 • Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, Dan Bikel, Lukas Blecher, Cristian Canton Ferrer, Moya Chen, Guillem Cucurull, David Esiobu, Jude Fernandes, Jeremy Fu, Wenyin Fu, Brian Fuller, Cynthia Gao, Vedanuj Goswami, Naman Goyal, Anthony Hartshorn, Saghar Hosseini, Rui Hou, Hakan Inan, Marcin Kardas, Viktor Kerkez, Madian Khabsa, Isabel Kloumann, Artem Korenev, Punit Singh Koura, Marie-Anne Lachaux, Thibaut Lavril, Jenya Lee, Diana Liskovich, Yinghai Lu, Yuning Mao, Xavier Martinet, Todor Mihaylov, Pushkar Mishra, Igor Molybog, Yixin Nie, Andrew Poulton, Jeremy Reizenstein, Rashi Rungta, Kalyan Saladi, Alan Schelten, Ruan Silva, Eric Michael Smith, Ranjan Subramanian, Xiaoqing Ellen Tan, Binh Tang, Ross Taylor, Adina Williams, Jian Xiang Kuan, Puxin Xu, Zheng Yan, Iliyan Zarov, Yuchen Zhang, Angela Fan, Melanie Kambadur, Sharan Narang, Aurelien Rodriguez, Robert Stojnic, Sergey Edunov, Thomas Scialom
In this work, we develop and release Llama 2, a collection of pretrained and fine-tuned large language models (LLMs) ranging in scale from 7 billion to 70 billion parameters.
Ranked #2 on Question Answering on PubChemQA
2 code implementations • 28 Nov 2022 • Tamara Czinczoll, Helen Yannakoudakis, Pushkar Mishra, Ekaterina Shutova
This paper examines the encoding of analogy in large-scale pretrained language models, such as BERT and GPT-2.
no code implementations • 20 Jul 2022 • Yihong Chen, Pushkar Mishra, Luca Franceschi, Pasquale Minervini, Pontus Stenetorp, Sebastian Riedel
Factorisation-based Models (FMs), such as DistMult, have enjoyed enduring success for Knowledge Graph Completion (KGC) tasks, often outperforming Graph Neural Networks (GNNs).
no code implementations • 27 Apr 2022 • David Adkins, Bilal Alsallakh, Adeel Cheema, Narine Kokhlikyan, Emily McReynolds, Pushkar Mishra, Chavez Procope, Jeremy Sawruk, Erin Wang, Polina Zvyagina
We further propose a preliminary approach, called Method Cards, which aims to increase the transparency and reproducibility of ML systems by providing prescriptive documentation of commonly-used ML methods and techniques.
1 code implementation • ACL 2021 • Rishav Hada, Sohi Sudhir, Pushkar Mishra, Helen Yannakoudakis, Saif M. Mohammad, Ekaterina Shutova
On social media platforms, hateful and offensive language negatively impact the mental well-being of users and the participation of people from diverse backgrounds.
no code implementations • Findings (EMNLP) 2021 • Pushkar Mishra, Helen Yannakoudakis, Ekaterina Shutova
Specifically, we review and analyze the state of the art methods that leverage user or community information to enhance the understanding and detection of abusive language.
1 code implementation • EACL 2021 • Niels van der Heijden, Helen Yannakoudakis, Pushkar Mishra, Ekaterina Shutova
The great majority of languages in the world are considered under-resourced for the successful application of deep learning methods.
Cross-Lingual Document Classification Document Classification +2
1 code implementation • 10 Sep 2020 • Nithin Holla, Pushkar Mishra, Helen Yannakoudakis, Ekaterina Shutova
Lifelong learning requires models that can continuously learn from sequential streams of data without suffering catastrophic forgetting due to shifts in data distributions.
1 code implementation • 14 Aug 2020 • Shantanu Chandra, Pushkar Mishra, Helen Yannakoudakis, Madhav Nimishakavi, Marzieh Saeidi, Ekaterina Shutova
Existing research has modeled the structure, style, content, and patterns in dissemination of online posts, as well as the demographic traits of users who interact with them.
no code implementations • ACL 2020 • Santhosh Rajamanickam, Pushkar Mishra, Helen Yannakoudakis, Ekaterina Shutova
The rise of online communication platforms has been accompanied by some undesirable effects, such as the proliferation of aggressive and abusive behaviour online.
2 code implementations • Findings of the Association for Computational Linguistics 2020 • Nithin Holla, Pushkar Mishra, Helen Yannakoudakis, Ekaterina Shutova
Meta-learning aims to solve this problem by training a model on a large number of few-shot tasks, with an objective to learn new tasks quickly from a small number of examples.
no code implementations • 17 Jan 2020 • Pushkar Mishra, Aleksandra Piktus, Gerard Goossen, Fabrizio Silvestri
Graph Neural Networks (GNNs) have received a lot of interest in the recent times.
no code implementations • 13 Aug 2019 • Pushkar Mishra, Helen Yannakoudakis, Ekaterina Shutova
Abuse on the Internet represents an important societal problem of our time.
no code implementations • SEMEVAL 2019 • Guy Aglionby, Chris Davis, Pushkar Mishra, Andrew Caines, Helen Yannakoudakis, Marek Rei, Ekaterina Shutova, Paula Buttery
We describe the CAMsterdam team entry to the SemEval-2019 Shared Task 6 on offensive language identification in Twitter data.
no code implementations • NAACL 2019 • Pushkar Mishra, Marco del Tredici, Helen Yannakoudakis, Ekaterina Shutova
Abuse on the Internet represents a significant societal problem of our time.
no code implementations • 14 Feb 2019 • Pushkar Mishra, Marco del Tredici, Helen Yannakoudakis, Ekaterina Shutova
The rapid growth of social media in recent years has fed into some highly undesirable phenomena such as proliferation of abusive and offensive language on the Internet.
no code implementations • WS 2018 • Pushkar Mishra, Helen Yannakoudakis, Ekaterina Shutova
The current state of the art approaches to abusive language detection, based on recurrent neural networks, do not explicitly address this problem and resort to a generic OOV (out of vocabulary) embedding for unseen words.
1 code implementation • COLING 2018 • Pushkar Mishra, Marco del Tredici, Helen Yannakoudakis, Ekaterina Shutova
The rapid growth of social media in recent years has fed into some highly undesirable phenomena such as proliferation of hateful and offensive language on the Internet.