no code implementations • 12 Sep 2024 • Ahmet Kapkiç, Pratanu Mandal, Shu Wan, Paras Sheth, Abhinav Gorantla, Yoonhyuk Choi, Huan Liu, K. Selçuk Candan
In this paper, we introduce {\em CausalBench}, a transparent, fair, and easy-to-use evaluation platform, aiming to (a) enable the advancement of research in causal learning by facilitating scientific collaboration in novel algorithms, datasets, and metrics and (b) promote scientific objectivity, reproducibility, fairness, and awareness of bias in causal learning research.
no code implementations • 17 Apr 2024 • Paras Sheth, Tharindu Kumarage, Raha Moraffah, Aman Chadha, Huan Liu
Content moderation faces a challenging task as social media's ability to spread hate speech contrasts with its role in promoting global connectivity.
1 code implementation • 19 Mar 2024 • Ayushi Nirmal, Amrita Bhattacharjee, Paras Sheth, Huan Liu
Although social media platforms are a prominent arena for users to engage in interpersonal discussions and express opinions, the facade and anonymity offered by social media may allow users to spew hate speech and offensive content.
no code implementations • 2 Mar 2024 • Tharindu Kumarage, Garima Agrawal, Paras Sheth, Raha Moraffah, Aman Chadha, Joshua Garland, Huan Liu
We have witnessed lately a rapid proliferation of advanced Large Language Models (LLMs) capable of generating high-quality text.
no code implementations • 5 Feb 2024 • Raha Moraffah, Paras Sheth, Saketh Vishnubhatla, Huan Liu
This survey focuses on the current study of causal feature selection: what it is and how it can reinforce the four aspects of responsible ML.
no code implementations • 8 Oct 2023 • Tharindu Kumarage, Paras Sheth, Raha Moraffah, Joshua Garland, Huan Liu
The novel universal evasive prompt is achieved in two steps: First, we create an evasive soft prompt tailored to a specific PLM through prompt tuning; and then, we leverage the transferability of soft prompts to transfer the learned evasive soft prompt from one PLM to another.
1 code implementation • 3 Aug 2023 • Paras Sheth, Tharindu Kumarage, Raha Moraffah, Aman Chadha, Huan Liu
By disentangling input into platform-dependent features (useful for predicting hate targets) and platform-independent features (used to predict the presence of hate), we learn invariant representations resistant to distribution shifts.
no code implementations • 25 Jul 2023 • Suraj Jyothi Unni, Paras Sheth, Kaize Ding, Huan Liu, K. Selcuk Candan
Discovering causal relationships in complex socio-behavioral systems is challenging but essential for informed decision-making.
1 code implementation • 10 Jul 2023 • Faisal Alatawi, Paras Sheth, Huan Liu
To facilitate measuring distances between users, we propose EchoGAE, a self-supervised graph autoencoder-based user embedding model that leverages users' posts and the interaction graph to embed them in a manner that reflects their ideological similarity.
1 code implementation • 15 Jun 2023 • Paras Sheth, Tharindu Kumarage, Raha Moraffah, Aman Chadha, Huan Liu
Hate speech detection refers to the task of detecting hateful content that aims at denigrating an individual or a group based on their religion, gender, sexual orientation, or other characteristics.
no code implementations • 30 Sep 2022 • Paras Sheth, Raha Moraffah, K. Selçuk Candan, Adrienne Raglin, Huan Liu
As a result models that rely on this assumption exhibit poor generalization capabilities.
no code implementations • 7 Aug 2022 • Mansooreh Karami, Ahmadreza Mosallanezhad, Paras Sheth, Huan Liu
To reduce the bias induced by the contributors, in this work, we focus on highlighting the engagers' contributions in the observed data as they are more likely to contribute when compared to lurkers, and they comprise a bigger population as compared to the contributors.
1 code implementation • 30 Jun 2022 • Bohan Jiang, Paras Sheth, Baoxin Li, Huan Liu
Despite the astonishing success of COVID-19 vaccines against the virus, a substantial proportion of the population is still hesitant to be vaccinated, undermining governmental efforts to control the virus.
no code implementations • 14 Apr 2022 • Paras Sheth, Ruocheng Guo, Lu Cheng, Huan Liu, K. Selçuk Candan
Aside from the user conformity, aspects of confounding such as item popularity present in the network information is also captured in our method with the aid of \textit{causal disentanglement} which unravels the learned representations into independent factors that are responsible for (a) modeling the exposure of an item to the user, (b) predicting the ratings, and (c) controlling the hidden confounders.
no code implementations • 7 Feb 2022 • Lu Cheng, Ruocheng Guo, Raha Moraffah, Paras Sheth, K. Selcuk Candan, Huan Liu
To bridge from conventional causal inference (i. e., based on statistical methods) to causal learning with big data (i. e., the intersection of causal inference and machine learning), in this survey, we review commonly-used datasets, evaluation methods, and measures for causal learning using an evaluation pipeline similar to conventional machine learning.
no code implementations • 25 Apr 2021 • Lu Cheng, Ahmadreza Mosallanezhad, Paras Sheth, Huan Liu
The goal of this survey is to bring forefront the potentials and promises of CL for SRAI.
no code implementations • 11 Feb 2021 • Raha Moraffah, Paras Sheth, Mansooreh Karami, Anchit Bhattacharya, Qianru Wang, Anique Tahir, Adrienne Raglin, Huan Liu
In this paper, we focus on two causal inference tasks, i. e., treatment effect estimation and causal discovery for time series data, and provide a comprehensive review of the approaches in each task.