no code implementations • 21 Jul 2024 • Gaurav Verma, Rynaa Grover, Jiawei Zhou, Binny Mathew, Jordan Kraemer, Munmun De Choudhury, Srijan Kumar
In contrast to prior work that has demonstrated the effectiveness of such classifiers in detecting hateful speech ($F_1 = 0. 89$), our work shows that accurate and reliable detection of violence-provoking speech is a challenging task ($F_1 = 0. 69$).
no code implementations • 22 Feb 2024 • Somnath Banerjee, Maulindu Sarkar, Punyajoy Saha, Binny Mathew, Animesh Mukherjee
Second, in a dataset extension exercise, using influence functions to automatically identify data points that have been initially `silver' annotated by some existing method and need to be cross-checked (and corrected) by annotators to improve the model performance.
1 code implementation • 6 May 2023 • Mithun Das, Rohit Raj, Punyajoy Saha, Binny Mathew, Manish Gupta, Animesh Mukherjee
Hate speech has become one of the most significant issues in modern society, having implications in both the online and the offline world.
Ranked #2 on
Hate Speech Detection
on HateMM
1 code implementation • 18 Mar 2023 • Punyajoy Saha, Kiran Garimella, Narla Komal Kalyan, Saurabh Kumar Pandey, Pauras Mangesh Meher, Binny Mathew, Animesh Mukherjee
Recently, social media platforms are heavily moderated to prevent the spread of online hate speech, which is usually fertile in toxic words and is directed toward an individual or a community.
no code implementations • 11 Feb 2023 • Piush Aggarwal, Pranit Chawla, Mithun Das, Punyajoy Saha, Binny Mathew, Torsten Zesch, Animesh Mukherjee
Empirically, we find a noticeable performance drop of as high as 10% in the macro-F1 score for certain attacks.
1 code implementation • 30 Nov 2022 • Punyajoy Saha, Divyanshu Sheth, Kushal Kedia, Binny Mathew, Animesh Mukherjee
We introduce two rationale-integrated BERT-based architectures (the RGFS models) and evaluate our systems over five different abusive language datasets, finding that in the few-shot classification setting, RGFS-based models outperform baseline models by about 7% in macro F1 scores and perform competitively to models finetuned on other source domains.
1 code implementation • 9 May 2022 • Punyajoy Saha, Kanishk Singh, Adarsh Kumar, Binny Mathew, Animesh Mukherjee
We generate counterspeech using three datasets and observe significant improvement across different attribute scores.
1 code implementation • LREC 2022 • Mithun Das, Punyajoy Saha, Binny Mathew, Animesh Mukherjee
To enable more targeted diagnostic insights of such multilingual hate speech models, we introduce a set of functionalities for the purpose of evaluation.
1 code implementation • 1 Aug 2021 • Mithun Das, Punyajoy Saha, Ritam Dutt, Pawan Goyal, Animesh Mukherjee, Binny Mathew
Hate speech is regarded as one of the crucial issues plaguing the online social media.
2 code implementations • 7 Feb 2021 • Punyajoy Saha, Binny Mathew, Kiran Garimella, Animesh Mukherjee
We observe that users writing fear speech messages use various events and symbols to create the illusion of fear among the reader about a target community.
6 code implementations • 18 Dec 2020 • Binny Mathew, Punyajoy Saha, Seid Muhie Yimam, Chris Biemann, Pawan Goyal, Animesh Mukherjee
We also observe that models, which utilize the human rationales for training, perform better in reducing unintended bias towards target communities.
Ranked #3 on
Hate Speech Detection
on HateXplain
3 code implementations • 14 Apr 2020 • Sai Saketh Aluru, Binny Mathew, Punyajoy Saha, Animesh Mukherjee
Hate speech detection is a challenging problem with most of the datasets available in only one language: English.
1 code implementation • 27 Jan 2020 • Binny Mathew, Sandipan Sikdar, Florian Lemmerich, Markus Strohmaier
We introduce POLAR - a framework that adds interpretability to pre-trained word embeddings via the adoption of semantic differentials.
1 code implementation • 27 Sep 2019 • Punyajoy Saha, Binny Mathew, Pawan Goyal, Animesh Mukherjee
In this paper, we present our machine learning model, HateMonitor, developed for Hate Speech and Offensive Content Identification in Indo-European Languages (HASOC), a shared task at FIRE 2019.
no code implementations • 10 Sep 2019 • Binny Mathew, Suman Kalyan Maity, Pawan Goyal, Animesh Mukherjee
Our system is also able to predict ~ 25% of the correct case of merges within the first month of the merge and ~ 40% of the cases within a year.
2 code implementations • 17 Dec 2018 • Punyajoy Saha, Binny Mathew, Pawan Goyal, Animesh Mukherjee
With the online proliferation of hate speech, there is an urgent need for systems that can detect such harmful content.
no code implementations • 6 Dec 2018 • Binny Mathew, Navish Kumar, Ravina, Pawan Goyal, Animesh Mukherjee
We also build a supervised model for classifying the hateful and counterspeech accounts on Twitter and obtain an F-score of 0. 77.
Social and Information Networks
no code implementations • 4 Dec 2018 • Binny Mathew, Ritam Dutt, Pawan Goyal, Animesh Mukherjee
The present online social media platform is afflicted with several issues, with hate speech being on the predominant forefront.
Social and Information Networks
no code implementations • 17 Nov 2018 • Binny Mathew, Ritam Dutt, Suman Kalyan Maity, Pawan Goyal, Animesh Mukherjee
In particular, we observe that the choice to post the question as anonymous is dependent on the user's perception of anonymity and they often choose to speak about depression, anxiety, social ties and personal issues under the guise of anonymity.
2 code implementations • Proceedings of the International AAAI Conference on Web and Social Media 2019 • Binny Mathew, Hardik Tharad, Subham Rajgaria, Prajwal Singhania, Suman Kalyan Maity, Pawan Goyal, Animesh Mukherje
In this paper, we create and release the first ever dataset for counterspeech using comments from YouTube.
Social and Information Networks
no code implementations • WS 2017 • Binny Mathew, Suman Kalyan Maity, Pratip Sarkar, Animesh Mukherjee, Pawan Goyal
Word senses are not static and may have temporal, spatial or corpus-specific scopes.