1 code implementation • 2 Oct 2024 • Punyajoy Saha, Abhilash Datta, Abhik Jana, Animesh Mukherjee
We evaluate two frameworks for generating counterspeech responses - vanilla and type-controlled prompts - across four large language models.
no code implementations • 27 Jun 2024 • Seid Muhie Yimam, Daryna Dementieva, Tim Fischer, Daniil Moskovskiy, Naquee Rizwan, Punyajoy Saha, Sarthak Roy, Martin Semmann, Alexander Panchenko, Chris Biemann, Animesh Mukherjee
Despite regulations imposed by nations and social media platforms, such as recent EU regulations targeting digital violence, abusive content persists as a significant challenge.
1 code implementation • 22 Mar 2024 • Punyajoy Saha, Aalok Agrawal, Abhik Jana, Chris Biemann, Animesh Mukherjee
In terms of prompting, we find that our proposed strategies help in improving counter speech generation across all the models.
no code implementations • 22 Feb 2024 • Somnath Banerjee, Maulindu Sarkar, Punyajoy Saha, Binny Mathew, Animesh Mukherjee
Second, in a dataset extension exercise, using influence functions to automatically identify data points that have been initially `silver' annotated by some existing method and need to be cross-checked (and corrected) by annotators to improve the model performance.
no code implementations • 19 Feb 2024 • Naquee Rizwan, Paramananda Bhaskar, Mithun Das, Swadhin Satyaprakash Majhi, Punyajoy Saha, Animesh Mukherjee
In this study, we aim to investigate the efficacy of these visual language models in handling intricate tasks such as hate meme detection.
1 code implementation • 11 Feb 2024 • Mithun Das, Saurabh Kumar Pandey, Shivansh Sethi, Punyajoy Saha, Animesh Mukherjee
With the rise of online abuse, the NLP community has begun investigating the use of neural architectures to generate counterspeech that can "counter" the vicious tone of such abusive speech and dilute/ameliorate their rippling effect over the social network.
no code implementations • 19 Oct 2023 • Sarthak Roy, Ashish Harshavardhan, Animesh Mukherjee, Punyajoy Saha
Recently efforts have been made by social media platforms as well as researchers to detect hateful or toxic language using large language models.
1 code implementation • 6 May 2023 • Mithun Das, Rohit Raj, Punyajoy Saha, Binny Mathew, Manish Gupta, Animesh Mukherjee
Hate speech has become one of the most significant issues in modern society, having implications in both the online and the offline world.
1 code implementation • 18 Mar 2023 • Punyajoy Saha, Kiran Garimella, Narla Komal Kalyan, Saurabh Kumar Pandey, Pauras Mangesh Meher, Binny Mathew, Animesh Mukherjee
Recently, social media platforms are heavily moderated to prevent the spread of online hate speech, which is usually fertile in toxic words and is directed toward an individual or a community.
no code implementations • 11 Feb 2023 • Piush Aggarwal, Pranit Chawla, Mithun Das, Punyajoy Saha, Binny Mathew, Torsten Zesch, Animesh Mukherjee
Empirically, we find a noticeable performance drop of as high as 10% in the macro-F1 score for certain attacks.
1 code implementation • 30 Nov 2022 • Punyajoy Saha, Divyanshu Sheth, Kushal Kedia, Binny Mathew, Animesh Mukherjee
We introduce two rationale-integrated BERT-based architectures (the RGFS models) and evaluate our systems over five different abusive language datasets, finding that in the few-shot classification setting, RGFS-based models outperform baseline models by about 7% in macro F1 scores and perform competitively to models finetuned on other source domains.
1 code implementation • 7 Oct 2022 • Mithun Das, Somnath Banerjee, Punyajoy Saha, Animesh Mukherjee
To overcome the existing research's limitations, in this study, we develop an annotated dataset of 10K Bengali posts consisting of 5K actual and 5K Romanized Bengali tweets.
no code implementations • TRAC (COLING) 2022 • Millon Madhur Das, Punyajoy Saha, Mithun Das
The proliferation of online hate speech has necessitated the creation of algorithms which can detect toxicity.
1 code implementation • 9 May 2022 • Punyajoy Saha, Kanishk Singh, Adarsh Kumar, Binny Mathew, Animesh Mukherjee
We generate counterspeech using three datasets and observe significant improvement across different attribute scores.
1 code implementation • LREC 2022 • Mithun Das, Punyajoy Saha, Binny Mathew, Animesh Mukherjee
To enable more targeted diagnostic insights of such multilingual hate speech models, we introduce a set of functionalities for the purpose of evaluation.
no code implementations • 27 Nov 2021 • Somnath Banerjee, Maulindu Sarkar, Nancy Agrawal, Punyajoy Saha, Mithun Das
Hate speech is considered to be one of the major issues currently plaguing online social media.
1 code implementation • 27 Nov 2021 • Mithun Das, Somnath Banerjee, Punyajoy Saha
In this FIRE 2021 shared task - "HASOC- Abusive and Threatening language detection in Urdu" the organizers propose an abusive language detection dataset in Urdu along with threatening language detection.
1 code implementation • 1 Aug 2021 • Mithun Das, Punyajoy Saha, Ritam Dutt, Pawan Goyal, Animesh Mukherjee, Binny Mathew
Hate speech is regarded as one of the crucial issues plaguing the online social media.
1 code implementation • EACL (DravidianLangTech) 2021 • Debjoy Saha, Naman Paharia, Debajit Chakraborty, Punyajoy Saha, Animesh Mukherjee
Social media often acts as breeding grounds for different forms of offensive content.
2 code implementations • 7 Feb 2021 • Punyajoy Saha, Binny Mathew, Kiran Garimella, Animesh Mukherjee
We observe that users writing fear speech messages use various events and symbols to create the illusion of fear among the reader about a target community.
6 code implementations • 18 Dec 2020 • Binny Mathew, Punyajoy Saha, Seid Muhie Yimam, Chris Biemann, Pawan Goyal, Animesh Mukherjee
We also observe that models, which utilize the human rationales for training, perform better in reducing unintended bias towards target communities.
Ranked #3 on Hate Speech Detection on HateXplain
3 code implementations • 14 Apr 2020 • Sai Saketh Aluru, Binny Mathew, Punyajoy Saha, Animesh Mukherjee
Hate speech detection is a challenging problem with most of the datasets available in only one language: English.
1 code implementation • 27 Sep 2019 • Punyajoy Saha, Binny Mathew, Pawan Goyal, Animesh Mukherjee
In this paper, we present our machine learning model, HateMonitor, developed for Hate Speech and Offensive Content Identification in Indo-European Languages (HASOC), a shared task at FIRE 2019.
2 code implementations • 17 Dec 2018 • Punyajoy Saha, Binny Mathew, Pawan Goyal, Animesh Mukherjee
With the online proliferation of hate speech, there is an urgent need for systems that can detect such harmful content.