Search Results for author: Binny Mathew

Found 20 papers, 13 papers with code

Deep Dive into Anonymity: A Large Scale Analysis of Quora Questions

no code implementations17 Nov 2018 Binny Mathew, Ritam Dutt, Suman Kalyan Maity, Pawan Goyal, Animesh Mukherjee

In particular, we observe that the choice to post the question as anonymous is dependent on the user's perception of anonymity and they often choose to speak about depression, anxiety, social ties and personal issues under the guise of anonymity.

Spread of hate speech in online social media

no code implementations4 Dec 2018 Binny Mathew, Ritam Dutt, Pawan Goyal, Animesh Mukherjee

The present online social media platform is afflicted with several issues, with hate speech being on the predominant forefront.

Social and Information Networks

Analyzing the hate and counter speech accounts on Twitter

no code implementations6 Dec 2018 Binny Mathew, Navish Kumar, Ravina, Pawan Goyal, Animesh Mukherjee

We also build a supervised model for classifying the hateful and counterspeech accounts on Twitter and obtain an F-score of 0. 77.

Social and Information Networks

Competing Topic Naming Conventions in Quora: Predicting Appropriate Topic Merges and Winning Topics from Millions of Topic Pairs

no code implementations10 Sep 2019 Binny Mathew, Suman Kalyan Maity, Pawan Goyal, Animesh Mukherjee

Our system is also able to predict ~ 25% of the correct case of merges within the first month of the merge and ~ 40% of the cases within a year.

Anomaly Detection TAG

HateMonitors: Language Agnostic Abuse Detection in Social Media

1 code implementation27 Sep 2019 Punyajoy Saha, Binny Mathew, Pawan Goyal, Animesh Mukherjee

In this paper, we present our machine learning model, HateMonitor, developed for Hate Speech and Offensive Content Identification in Indo-European Languages (HASOC), a shared task at FIRE 2019.

Abuse Detection Abusive Language +1

The POLAR Framework: Polar Opposites Enable Interpretability of Pre-Trained Word Embeddings

1 code implementation27 Jan 2020 Binny Mathew, Sandipan Sikdar, Florian Lemmerich, Markus Strohmaier

We introduce POLAR - a framework that adds interpretability to pre-trained word embeddings via the adoption of semantic differentials.

Word Embeddings

HateXplain: A Benchmark Dataset for Explainable Hate Speech Detection

6 code implementations18 Dec 2020 Binny Mathew, Punyajoy Saha, Seid Muhie Yimam, Chris Biemann, Pawan Goyal, Animesh Mukherjee

We also observe that models, which utilize the human rationales for training, perform better in reducing unintended bias towards target communities.

Hate Speech Detection Text Classification

"Short is the Road that Leads from Fear to Hate": Fear Speech in Indian WhatsApp Groups

2 code implementations7 Feb 2021 Punyajoy Saha, Binny Mathew, Kiran Garimella, Animesh Mukherjee

We observe that users writing fear speech messages use various events and symbols to create the illusion of fear among the reader about a target community.

HateCheckHIn: Evaluating Hindi Hate Speech Detection Models

1 code implementation LREC 2022 Mithun Das, Punyajoy Saha, Binny Mathew, Animesh Mukherjee

To enable more targeted diagnostic insights of such multilingual hate speech models, we introduce a set of functionalities for the purpose of evaluation.

Hate Speech Detection

CounterGeDi: A controllable approach to generate polite, detoxified and emotional counterspeech

1 code implementation9 May 2022 Punyajoy Saha, Kanishk Singh, Adarsh Kumar, Binny Mathew, Animesh Mukherjee

We generate counterspeech using three datasets and observe significant improvement across different attribute scores.

Attribute

Rationale-Guided Few-Shot Classification to Detect Abusive Language

1 code implementation30 Nov 2022 Punyajoy Saha, Divyanshu Sheth, Kushal Kedia, Binny Mathew, Animesh Mukherjee

We introduce two rationale-integrated BERT-based architectures (the RGFS models) and evaluate our systems over five different abusive language datasets, finding that in the few-shot classification setting, RGFS-based models outperform baseline models by about 7% in macro F1 scores and perform competitively to models finetuned on other source domains.

Abusive Language Classification +1

HateProof: Are Hateful Meme Detection Systems really Robust?

no code implementations11 Feb 2023 Piush Aggarwal, Pranit Chawla, Mithun Das, Punyajoy Saha, Binny Mathew, Torsten Zesch, Animesh Mukherjee

Empirically, we find a noticeable performance drop of as high as 10% in the macro-F1 score for certain attacks.

Contrastive Learning

On the rise of fear speech in online social media

1 code implementation18 Mar 2023 Punyajoy Saha, Kiran Garimella, Narla Komal Kalyan, Saurabh Kumar Pandey, Pauras Mangesh Meher, Binny Mathew, Animesh Mukherjee

Recently, social media platforms are heavily moderated to prevent the spread of online hate speech, which is usually fertile in toxic words and is directed toward an individual or a community.

HateMM: A Multi-Modal Dataset for Hate Video Classification

1 code implementation6 May 2023 Mithun Das, Rohit Raj, Punyajoy Saha, Binny Mathew, Manish Gupta, Animesh Mukherjee

Hate speech has become one of the most significant issues in modern society, having implications in both the online and the offline world.

Classification Hate Speech Detection +1

InfFeed: Influence Functions as a Feedback to Improve the Performance of Subjective Tasks

no code implementations22 Feb 2024 Somnath Banerjee, Maulindu Sarkar, Punyajoy Saha, Binny Mathew, Animesh Mukherjee

Second, in a dataset extension exercise, using influence functions to automatically identify data points that have been initially `silver' annotated by some existing method and need to be cross-checked (and corrected) by annotators to improve the model performance.

Sarcasm Detection Stance Classification

Cannot find the paper you are looking for? You can Submit a new open access paper.