no code implementations • EMNLP 2020 • Adaku Uchendu, Thai Le, Kai Shu, Dongwon Lee
In recent years, the task of generating realistic short and long texts have made tremendous advancements.
1 code implementation • 25 Feb 2025 • Xiongxiao Xu, Haoran Wang, Yueqing Liang, Philip S. Yu, Yue Zhao, Kai Shu
Starting with the univariate case (point- and range-wise anomalies), we extend our evaluation to more practical scenarios, including multivariate and irregular time series scenarios, and variate-wise anomalies.
1 code implementation • 19 Feb 2025 • Yueqing Liang, Liangwei Yang, Chen Wang, Congying Xia, Rui Meng, Xiongxiao Xu, Haoran Wang, Ali Payani, Kai Shu
Large Language Models (LLMs) have achieved significant advances in natural language processing, yet their potential for high-stake political decision-making remains largely unexplored.
1 code implementation • 18 Feb 2025 • Yunpeng Xiao, Youpeng Zhao, Kai Shu
Some tasks, such as stance detection and sentiment analysis, are closely related to individual subjective perspectives, thus termed individual-level NLU.
no code implementations • 7 Dec 2024 • Junwei Yin, Min Gao, Kai Shu, Wentao Li, Yinqiu Huang, Zongwei Wang
To tackle these issues, we propose BREAK, a broad-range semantics model for fake news detection that leverages a fully connected graph to capture comprehensive semantics while employing dual denoising modules to minimize both structural and feature noise.
2 code implementations • 6 Dec 2024 • Kaustubh D. Dhole, Kai Shu, Eugene Agichtein
To validate the proposed techniques, we introduce ConQRet, a new benchmark featuring long and complex human-authored arguments on debated topics, grounded in real-world websites, allowing an exhaustive evaluation across retrieval effectiveness, argument quality, and groundedness.
1 code implementation • 25 Nov 2024 • Dawei Li, Bohan Jiang, Liangjie Huang, Alimohammad Beigi, Chengshuai Zhao, Zhen Tan, Amrita Bhattacharjee, YuXuan Jiang, Canyu Chen, Tianhao Wu, Kai Shu, Lu Cheng, Huan Liu
Assessment and evaluation have long been critical challenges in artificial intelligence (AI) and natural language processing (NLP).
no code implementations • 14 Nov 2024 • Haoran Wang, Aman Rangapur, Xiongxiao Xu, Yueqing Liang, Haroon Gharwi, Carl Yang, Kai Shu
Existing claim verification datasets often do not require systems to perform complex reasoning or effectively interpret multimodal evidence.
no code implementations • 10 Nov 2024 • Canyu Chen, Jian Yu, Shan Chen, Che Liu, Zhongwei Wan, Danielle Bitterman, Fei Wang, Kai Shu
Large Language Models (LLMs) hold great promise to revolutionize current clinical systems for their superior capacities on medical text processing tasks and medical licensing exams.
1 code implementation • 21 Oct 2024 • Baixiang Huang, Canyu Chen, Xiongxiao Xu, Ali Payani, Kai Shu
Large Language Models (LLMs) suffer from hallucinations, referring to the non-factual information in generated content, despite their superior capacities across tasks.
1 code implementation • 3 Oct 2024 • Xiongxiao Xu, Solomon Abera Bekele, Brice Videau, Kai Shu
Specifically, the proposed framework EnergyUCB (1) balances the performance-energy trade-off in the reward function, (2) effectively navigates the exploration & exploitation dilemma when adjusting GPU core frequencies online, and (3) leverages the ratio of GPU core utilization to uncore utilization as a real-time GPU performance metric.
no code implementations • 6 Sep 2024 • Kai Shu, Yuzhuo Jia, Ziyang Zhang, Jiechao Gao
Automatic Medical Imaging Narrative generation aims to alleviate the workload of radiologists by producing accurate clinical descriptions directly from radiological images.
no code implementations • 31 Jul 2024 • Alimohammad Beigi, Zhen Tan, Nivedh Mudiam, Canyu Chen, Kai Shu, Huan Liu
To address this, we introduce a novel approach based on Supervised Contrastive Learning.
1 code implementation • 29 Jul 2024 • Canyu Chen, Baixiang Huang, Zekun Li, Zhaorun Chen, Shiyang Lai, Xiongxiao Xu, Jia-Chen Gu, Jindong Gu, Huaxiu Yao, Chaowei Xiao, Xifeng Yan, William Yang Wang, Philip Torr, Dawn Song, Kai Shu
Then, we find that editing attacks can inject both types of misinformation into LLMs, and the effectiveness is particularly high for commonsense misinformation injection.
1 code implementation • 20 Jun 2024 • Yueqing Liang, Liangwei Yang, Chen Wang, Xiongxiao Xu, Philip S. Yu, Kai Shu
With the emergence of large language models (LLMs) and their ability to perform a variety of tasks, their application in recommender systems (RecSys) has shown promise.
no code implementations • 29 May 2024 • Qin Yang, Meisam Mohammad, Han Wang, Ali Payani, Ashish Kundu, Kai Shu, Yan Yan, Yuan Hong
To address such limitations, we propose a novel Language Model-based Optimal Differential Privacy (LMO-DP) mechanism, which takes the first step to enable the tight composition of accurately fine-tuning (large) language models with a sub-optimal DP mechanism, even in strong privacy regimes (e. g., $0. 1\leq \epsilon<3$).
2 code implementations • 23 Apr 2024 • Xiongxiao Xu, Canyu Chen, Yueqing Liang, Baixiang Huang, Guangji Bai, Liang Zhao, Kai Shu
To meet the objectives, we propose a multi-scale hybrid Mamba-Transformer experts model State Space Transformer (SST).
no code implementations • 14 Mar 2024 • Guanghua Li, Wensheng Lu, Wei zhang, Defu Lian, Kezhong Lu, Rui Mao, Kai Shu, Hao Liao
The proliferation of fake news has had far-reaching implications on politics, the economy, and society at large.
1 code implementation • 13 Mar 2024 • Baixiang Huang, Canyu Chen, Kai Shu
(3) Can LLMs provide explainability in authorship analysis, particularly through the role of linguistic features?
1 code implementation • 7 Feb 2024 • Chengxing Xie, Canyu Chen, Feiran Jia, Ziyu Ye, Shiyang Lai, Kai Shu, Jindong Gu, Adel Bibi, Ziniu Hu, David Jurgens, James Evans, Philip Torr, Bernard Ghanem, Guohao Li
In this paper, we focus on one critical and elemental behavior in human interactions, trust, and investigate whether LLM agents can simulate human trust behavior.
no code implementations • 27 Jan 2024 • Jianfei Xiao, Yancan Chen, Yimin Ou, Hanyi Yu, Kai Shu, Yiyong Xiao
Nevertheless, for the dialogue summarization task, which aims to generate summaries for different roles in dialogue, most of the state-of-the-art methods conduct on small models (e. g Bart and Bert).
1 code implementation • 10 Jan 2024 • Yue Huang, Lichao Sun, Haoran Wang, Siyuan Wu, Qihui Zhang, Yuan Li, Chujie Gao, Yixin Huang, Wenhan Lyu, Yixuan Zhang, Xiner Li, Zhengliang Liu, Yixin Liu, Yijue Wang, Zhikun Zhang, Bertie Vidgen, Bhavya Kailkhura, Caiming Xiong, Chaowei Xiao, Chunyuan Li, Eric Xing, Furong Huang, Hao liu, Heng Ji, Hongyi Wang, huan zhang, Huaxiu Yao, Manolis Kellis, Marinka Zitnik, Meng Jiang, Mohit Bansal, James Zou, Jian Pei, Jian Liu, Jianfeng Gao, Jiawei Han, Jieyu Zhao, Jiliang Tang, Jindong Wang, Joaquin Vanschoren, John Mitchell, Kai Shu, Kaidi Xu, Kai-Wei Chang, Lifang He, Lifu Huang, Michael Backes, Neil Zhenqiang Gong, Philip S. Yu, Pin-Yu Chen, Quanquan Gu, ran Xu, Rex Ying, Shuiwang Ji, Suman Jana, Tianlong Chen, Tianming Liu, Tianyi Zhou, William Wang, Xiang Li, Xiangliang Zhang, Xiao Wang, Xing Xie, Xun Chen, Xuyu Wang, Yan Liu, Yanfang Ye, Yinzhi Cao, Yong Chen, Yue Zhao
This paper introduces TrustLLM, a comprehensive study of trustworthiness in LLMs, including principles for different dimensions of trustworthiness, established benchmark, evaluation, and analysis of trustworthiness for mainstream LLMs, and discussion of open challenges and future directions.
no code implementations • 20 Nov 2023 • YiFan Li, Zhen Tan, Kai Shu, Zongsheng Cao, Yu Kong, Huan Liu
Graph Neural Networks (GNNs) have emerged as a powerful tool for representation learning on graphs, but they often suffer from overfitting and label noise issues, especially when the data is scarce or imbalanced.
1 code implementation • 15 Nov 2023 • Haoran Wang, Kai Shu
To ensure AI safety, instruction-tuned Large Language Models (LLMs) are specifically trained to ensure alignment, which refers to making models behave in accordance with human intentions.
no code implementations • 15 Nov 2023 • Yueqing Liang, Lu Cheng, Ali Payani, Kai Shu
This work investigates the potential of undermining both fairness and detection performance in abusive language detection.
1 code implementation • 8 Oct 2023 • Haoran Wang, Kai Shu
While existing works on claim verification have shown promising results, a crucial piece of the puzzle that remains unsolved is to understand how to verify claims without relying on human-annotated data, which is expensive to create at a large scale.
1 code implementation • 25 Sep 2023 • Canyu Chen, Kai Shu
Then, through extensive empirical investigation, we discover that LLM-generated misinformation can be harder to detect for humans and detectors compared to human-written misinformation with the same semantics, which suggests it can have more deceptive styles and potentially cause more harm.
1 code implementation • 15 Sep 2023 • Aman Rangapur, Haoran Wang, Ling Jian, Kai Shu
Fact-checking in financial domain is under explored, and there is a shortage of quality dataset in this domain.
no code implementations • 6 Sep 2023 • Aman Rangapur, Haoran Wang, Kai Shu
In conclusion, this research paper sheds light on the pervasive issue of online financial misinformation and its wide-ranging consequences.
no code implementations • 27 Jun 2023 • Junwei Yin, Min Gao, Kai Shu, Zehua Zhao, Yinqiu Huang, Jia Wang
To this end, we propose an approach of Emulating the behaviors of readers (Ember) for fake news detection on social media, incorporating readers' reading and verificating process to model news from the component perspective thoroughly.
no code implementations • 14 Jun 2023 • Kai Shu, Le Wu, Yuchang Zhao, Aiping Liu, Ruobing Qian, Xun Chen
The randomness of input noise and the precise representation enable the synthetic samples to possess diversity while ensuring the consistency of feature space.
1 code implementation • 18 May 2023 • Xiongxiao Xu, Kaize Ding, Canyu Chen, Kai Shu
However, the work exploring limited labeled anomalies and a large amount of unlabeled nodes in graphs to detect anomalies is relatively limited.
1 code implementation • 17 Apr 2023 • Baixiang Huang, Bryan Hooi, Kai Shu
To bridge this gap, we have constructed a real-world graph-based Traffic Accident Prediction (TAP) data repository, along with two representative tasks: accident occurrence prediction and accident severity prediction.
1 code implementation • 24 Dec 2022 • Ujun Jeong, Kaize Ding, Lu Cheng, Ruocheng Guo, Kai Shu, Huan Liu
Nowadays, fake news easily propagates through online social networks and becomes a grand threat to individuals and society.
Ranked #1 on
Graph Classification
on UPFD-POL
no code implementations • 10 Nov 2022 • Canyu Chen, Haoran Wang, Matthew Shapiro, Yunyu Xiao, Fei Wang, Kai Shu
Because of the uniqueness and importance of combating health misinformation in social media, we conduct this survey to further facilitate interdisciplinary research on this problem.
no code implementations • 18 Jul 2022 • Canyu Chen, Yueqing Liang, Xiongxiao Xu, Shangyu Xie, Ashish Kundu, Ali Payani, Yuan Hong, Kai Shu
Thus, it is essential to ensure fairness in machine learning models.
1 code implementation • 26 Jun 2022 • Yongchun Zhu, Qiang Sheng, Juan Cao, Qiong Nan, Kai Shu, Minghui Wu, Jindong Wang, Fuzhen Zhuang
In this paper, we propose a Memory-guided Multi-view Multi-domain Fake News Detection Framework (M$^3$FEND) to address these two challenges.
2 code implementations • 21 Jun 2022 • Kay Liu, Yingtong Dou, Yue Zhao, Xueying Ding, Xiyang Hu, Ruitong Zhang, Kaize Ding, Canyu Chen, Hao Peng, Kai Shu, Lichao Sun, Jundong Li, George H. Chen, Zhihao Jia, Philip S. Yu
To bridge this gap, we present--to the best of our knowledge--the first comprehensive benchmark for unsupervised outlier node detection on static attributed graphs called BOND, with the following highlights.
1 code implementation • 8 Jun 2022 • Yueqing Liang, Canyu Chen, Tian Tian, Kai Shu
Though we lack the sensitive attribute for training a fair model in the target domain, there might exist a similar domain that has sensitive attributes.
1 code implementation • 18 May 2022 • Canyu Chen, Kai Shu
Extensive experiment results on few-shot text classification tasks demonstrate the superior performance of the proposed framework by effectively leveraging label semantics and data augmentation for natural language understanding.
1 code implementation • 6 May 2022 • Qiang Sheng, Juan Cao, H. Russell Bernard, Kai Shu, Jintao Li, Huan Liu
False news that spreads on social media has proliferated over the past years and has led to multi-aspect threats in the real world.
no code implementations • 16 Feb 2022 • Ahmadreza Mosallanezhad, Mansooreh Karami, Kai Shu, Michelle V. Mancenido, Huan Liu
With social media being a major force in information consumption, accelerated propagation of fake news has presented new challenges for platforms to distinguish between legitimate and fake news.
no code implementations • 9 Feb 2022 • Xinyi Zhou, Kai Shu, Vir V. Phoha, Huan Liu, Reza Zafarani
To distinguish between intentional versus unintentional spreading, we study the psychological explanations of unintentional spreading.
1 code implementation • EMNLP (CINLP) 2021 • Zhao Wang, Kai Shu, Aron Culotta
In this paper, we propose a simple and intuitive regularization approach to integrate causal knowledge during model training and build a robust and fair model by emphasizing causal features and de-emphasizing spurious features.
no code implementations • NAACL 2022 • Guoqing Zheng, Giannis Karamanolakis, Kai Shu, Ahmed Hassan Awadallah
In this paper, we propose such a benchmark, named WALNUT (semi-WeAkly supervised Learning for Natural language Understanding Testbed), to advocate and facilitate research on weak supervision for NLU.
no code implementations • 8 Jun 2021 • Enyan Dai, Kai Shu, Yiwei Sun, Suhang Wang
We propose a novel generative framework named as ADDES which can synthesize high-quality labeled data for target classification tasks by learning from data with inexact supervision and the relations between inexact supervision and target classes.
1 code implementation • 29 Apr 2021 • Tianxiang Zhao, Enyan Dai, Kai Shu, Suhang Wang
Though the sensitive attribute of each data sample is unknown, we observe that there are usually some non-sensitive features in the training data that are highly correlated with sensitive attributes, which can be used to alleviate the bias.
2 code implementations • 25 Apr 2021 • Yingtong Dou, Kai Shu, Congying Xia, Philip S. Yu, Lichao Sun
The majority of existing fake news detection algorithms focus on mining news content and/or the surrounding exogenous context for discovering deceptive signals; while the endogenous preference of a user when he/she decides to spread a piece of fake news or not is ignored.
Ranked #1 on
Graph Classification
on UPFD-GOS
1 code implementation • 8 Dec 2020 • Kai Shu, Yichuan Li, Kaize Ding, Huan Liu
The existing text generation methods either afford limited supplementary information or lose consistency between the input and output which makes the synthetic news less trustworthy.
3 code implementations • 8 Nov 2020 • Yichuan Li, Bohan Jiang, Kai Shu, Huan Liu
The COVID-19 epidemic is considered as the global health crisis of the whole society and the greatest challenge mankind faced since World War Two.
Social and Information Networks Computers and Society
no code implementations • 3 Nov 2020 • Hao Liao, Qixin Liu, Kai Shu, Xing Xie
Yet, the popularity of social media also provides opportunities to better detect fake news.
Fake News Detection
Representation Learning
Social and Information Networks
no code implementations • 30 Oct 2020 • Ahmadreza Mosallanezhad, Kai Shu, Huan Liu
In this paper, we consider realistic news as news that cannot be easily detected by a fake news classifier.
no code implementations • 18 Oct 2020 • Amrita Bhattacharjee, Kai Shu, Min Gao, Huan Liu
We then proceed to discuss the inherent challenges in disinformation research, and then elaborate on the computational and interdisciplinary approaches towards mitigation of disinformation, after a short overview of the various directions explored in detection efforts.
no code implementations • 14 Jul 2020 • Kai Shu, Amrita Bhattacharjee, Faisal Alatawi, Tahora Nazer, Kaize Ding, Mansooreh Karami, Huan Liu
The creation, dissemination, and consumption of disinformation and fabricated content on social media is a growing concern, especially with the ease of access to such sources, and the lack of awareness of the existence of such false information.
2 code implementations • 23 Jun 2020 • Kaize Ding, Jianling Wang, Jundong Li, Kai Shu, Chenghao Liu, Huan Liu
By constructing a pool of semi-supervised node classification tasks to mimic the real test environment, GPN is able to perform \textit{meta-learning} on an attributed network and derive a highly generalizable model for handling the target classification task.
no code implementations • 26 May 2020 • Kai Shu, Subhabrata Mukherjee, Guoqing Zheng, Ahmed Hassan Awadallah, Milad Shokouhi, Susan Dumais
In this paper, we propose to leverage user actions as a source of weak supervision, in addition to a limited set of annotated examples, to detect intents in emails.
no code implementations • 3 Apr 2020 • Kai Shu, Guoqing Zheng, Yichuan Li, Subhabrata Mukherjee, Ahmed Hassan Awadallah, Scott Ruston, Huan Liu
Social media has greatly enabled people to participate in online activities at an unprecedented rate.
1 code implementation • 2 Jan 2020 • Kai Shu, Suhang Wang, Dongwon Lee, Huan Liu
In recent years, disinformation including fake news, has became a global phenomenon due to its explosive growth, particularly on social media.
no code implementations • 28 Oct 2019 • Raha Moraffah, Kai Shu, Adrienne Raglin, Huan Liu
Recent research on deep domain adaptation proposed to mitigate this problem by forcing the deep model to learn more transferable feature representations across domains.
no code implementations • 24 Oct 2019 • Kai Shu, Ahmed Hassan Awadallah, Susan Dumais, Huan Liu
This is especially the case for many real-world tasks where large scale annotated examples are either too expensive to acquire or unavailable due to privacy or data access constraints.
no code implementations • 1 May 2019 • Yu-Feng Yu, Yuelong Zhu, Dingsheng Wan, Qun Zhao, Kai Shu, Huan Liu
Floods of research and practical applications employ social media data for a wide range of public applications, including environmental monitoring, water resource managing, disaster and emergency response. Hydroinformatics can benefit from the social media technologies with newly emerged data, techniques and analytical tools to handle large datasets, from which creative ideas and new values could be mined. This paper first proposes a 4W (What, Why, When, hoW) model and a methodological structure to better understand and represent the application of social media to hydroinformatics, then provides an overview of academic research of applying social media to hydroinformatics such as water environment, water resources, flood, drought and water Scarcity management.
no code implementations • 30 Apr 2019 • Kai Shu, Xinyi Zhou, Suhang Wang, Reza Zafarani, Huan Liu
In an attempt to understand connections between user profiles and fake news, first, we measure users' sharing behaviors on social media and group representative users who are more likely to share fake and real news; then, we perform a comparative analysis of explicit and implicit profile features between these user groups, which reveals their potential to help differentiate fake news from real news.
2 code implementations • 21 Mar 2019 • Kai Shu, Deepak Mahudeswaran, Suhang Wang, Huan Liu
In an attempt to understand the correlations between news propagation networks and fake news, first, we build a hierarchical propagation network from macro-level and micro-level of fake news and true news; second, we perform a comparative analysis of the propagation network features of linguistic, structural and temporal perspectives between fake and real news, which demonstrates the potential of utilizing these features to detect fake news; third, we show the effectiveness of these propagation network features for fake news detection.
Social and Information Networks
no code implementations • 6 Mar 2019 • Wen Zhang, Kai Shu, Huan Liu, Yalin Wang
In particular, we provide a principled approach to jointly capture local and global information in the user-user social graph and propose the framework {\m}, which jointly learning user representations for user identity linkage.
1 code implementation • 5 Mar 2019 • Xueyao Zhang, Juan Cao, Xirong Li, Qiang Sheng, Lei Zhong, Kai Shu
Emotion plays an important role in detecting fake news online.
9 code implementations • 5 Sep 2018 • Kai Shu, Deepak Mahudeswaran, Suhang Wang, Dongwon Lee, Huan Liu
However, fake news detection is a non-trivial task, which requires multi-source information such as news content, social context, and dynamic information.
Social and Information Networks
4 code implementations • 20 Dec 2017 • Kai Shu, Suhang Wang, Huan Liu
Recent Social and Psychology studies show potential importance to utilize social media data: 1) Confirmation bias effect reveals that consumers prefer to believe information that confirms their existing stances; 2) Echo chamber effect suggests that people tend to follow likeminded users and form segregated communities on social media.
Social and Information Networks
no code implementations • 14 Sep 2017 • Fred Morstatter, Kai Shu, Suhang Wang, Huan Liu
We apply our solution to sentiment analysis, a task that can benefit from the emoji calibration technique we use in this work.
9 code implementations • 7 Aug 2017 • Kai Shu, Amy Sliva, Suhang Wang, Jiliang Tang, Huan Liu
First, fake news is intentionally written to mislead readers to believe false information, which makes it difficult and nontrivial to detect based on news content; therefore, we need to include auxiliary information, such as user social engagements on social media, to help make a determination.