no code implementations • 19 Aug 2024 • Jingyu Hu, Weiru Liu, Mengnan Du
Recent studies highlight the effectiveness of using in-context learning (ICL) to steer large language models (LLMs) in processing tabular data, a challenging task given the structured nature of such data.
no code implementations • 27 Jul 2024 • Dong Shu, Haoran Zhao, Xukun Liu, David Demeter, Mengnan Du, Yongfeng Zhang
Moreover, the subtle distinctions between similar and precedent cases require a deep understanding of legal knowledge.
1 code implementation • 15 Jul 2024 • Chong Zhang, Xinyi Liu, Mingyu Jin, Zhongmou Zhang, Lingyao Li, Zhenting Wang, Wenyue Hua, Dong Shu, Suiyuan Zhu, Xiaobo Jin, Sujian Li, Mengnan Du, Yongfeng Zhang
The StockAgent allows users to evaluate the impact of different external factors on investor trading and to analyze trading behavior and profitability effects.
no code implementations • 28 Jun 2024 • Haoyi Xiong, Jiang Bian, Yuchen Li, Xuhong LI, Mengnan Du, Shuaiqiang Wang, Dawei Yin, Sumi Helal
Combining Large Language Models (LLMs) with search engine services marks a significant shift in the field of services computing, opening up new possibilities to enhance how we search for and retrieve information, understand content, and interact with internet services.
1 code implementation • 23 May 2024 • Hanrong Zhang, Zhenting Wang, Tingxu Han, Mingyu Jin, Chenlu Zhan, Mengnan Du, Hongwei Wang, Shiqing Ma
In this paper, we propose an imperceptible and effective backdoor attack against self-supervised models.
no code implementations • 21 May 2024 • Yutao Du, Qin Li, Raghav Gnanasambandam, Mengnan Du, Haimin Wang, Bo Shen
Studying the sun's outer atmosphere is challenging due to its complex magnetic fields impacting solar activities.
no code implementations • 17 Apr 2024 • Zihao Li, Yucheng Shi, Zirui Liu, Fan Yang, Ali Payani, Ninghao Liu, Mengnan Du
Besides, the experiments show that there is a strong correlation between the LLM's performance in different languages and the proportion of those languages in its pre-training corpus.
1 code implementation • 10 Apr 2024 • Mingyu Jin, Qinkai Yu, Jingyuan Huang, Qingcheng Zeng, Zhenting Wang, Wenyue Hua, Haiyan Zhao, Kai Mei, Yanda Meng, Kaize Ding, Fan Yang, Mengnan Du, Yongfeng Zhang
In this paper, we explore the hypothesis that LLMs process concepts of varying complexities in different layers, introducing the idea of "Concept Depth" to suggest that more complex concepts are typically acquired in deeper layers.
1 code implementation • 30 Mar 2024 • Mingyu Jin, Haochen Xue, Zhenting Wang, Boming Kang, Ruosong Ye, Kaixiong Zhou, Mengnan Du, Yongfeng Zhang
Specifically, we propose Protein Chain of Thought (ProCoT), which replicates the biological mechanism of signaling pathways as natural language prompts.
1 code implementation • 13 Mar 2024 • Xuansheng Wu, Haiyan Zhao, Yaochen Zhu, Yucheng Shi, Fan Yang, Tianming Liu, Xiaoming Zhai, Wenlin Yao, Jundong Li, Mengnan Du, Ninghao Liu
Therefore, in this paper, we introduce Usable XAI in the context of LLMs by analyzing (1) how XAI can benefit LLMs and AI systems, and (2) how LLMs can contribute to the advancement of XAI.
no code implementations • 12 Mar 2024 • Dong Shu, Tianle Chen, Mingyu Jin, Chong Zhang, Mengnan Du, Yongfeng Zhang
We first convert structured knowledge graph data into natural language and then use these natural language prompts to fine-tune LLMs to enhance multi-hop link prediction in KGs.
1 code implementation • 20 Feb 2024 • Mingyu Jin, Beichen Wang, Zhaoqian Xue, Suiyuan Zhu, Wenyue Hua, Hua Tang, Kai Mei, Mengnan Du, Yongfeng Zhang
In this study, we introduce "CosmoAgent," an innovative artificial intelligence framework utilizing Large Language Models (LLMs) to simulate complex interactions between human and extraterrestrial civilizations, with a special emphasis on Stephen Hawking's cautionary advice about not sending radio signals haphazardly into the universe.
no code implementations • 16 Feb 2024 • Haiyan Zhao, Fan Yang, Bo Shen, Himabindu Lakkaraju, Mengnan Du
Large language models (LLMs) have led to breakthroughs in language tasks, yet the internal mechanisms that enable their remarkable generalization and reasoning abilities remain opaque.
no code implementations • 16 Feb 2024 • Hua Tang, Chong Zhang, Mingyu Jin, Qinkai Yu, Zhenting Wang, Xiaobo Jin, Yongfeng Zhang, Mengnan Du
Large language models (LLMs) have been applied in many fields and have developed rapidly in recent years.
no code implementations • 7 Feb 2024 • Yu-Neng Chuang, Guanchu Wang, Chia-Yuan Chang, Ruixiang Tang, Shaochen Zhong, Fan Yang, Mengnan Du, Xuanting Cai, Xia Hu
Large Language Models (LLMs) have become proficient in addressing complex tasks by leveraging their extensive internal knowledge and reasoning capabilities.
no code implementations • 1 Feb 2024 • Mingyu Jin, Qinkai Yu, Dong Shu, Chong Zhang, Lizhou Fan, Wenyue Hua, Suiyuan Zhu, Yanda Meng, Zhenting Wang, Mengnan Du, Yongfeng Zhang
Compared to traditional health management applications, our system has three main advantages: (1) It integrates health reports and medical knowledge into a large model to ask relevant questions to large language model for disease prediction; (2) It leverages a retrieval augmented generation (RAG) mechanism to enhance feature extraction; (3) It incorporates a semi-automated feature updating framework that can merge and delete features to improve accuracy of disease prediction.
no code implementations • 27 Jan 2024 • Junyi Ye, Mengnan Du, Guiling Wang
This paper introduces DataFrame question answering (QA), a novel task that utilizes large language models (LLMs) to generate Pandas queries for information retrieval and data analysis on dataframes, emphasizing safe and non-revealing data handling.
no code implementations • 16 Jan 2024 • Jiamin Chen, Xuhong LI, Yanwu Xu, Mengnan Du, Haoyi Xiong
Based on a large-scale medical image classification dataset, our work collects explanations from well-trained classifiers to generate pseudo labels of segmentation tasks.
1 code implementation • 16 Jan 2024 • Zichuan Liu, Yingying Zhang, Tianchun Wang, Zefan Wang, Dongsheng Luo, Mengnan Du, Min Wu, Yi Wang, Chunlin Chen, Lunting Fan, Qingsong Wen
Explaining multivariate time series is a compound challenge, as it requires identifying important locations in the time series and matching complex temporal patterns.
2 code implementations • 10 Jan 2024 • Mingyu Jin, Qinkai Yu, Dong Shu, Haiyan Zhao, Wenyue Hua, Yanda Meng, Yongfeng Zhang, Mengnan Du
Alternatively, shortening the reasoning steps, even while preserving the key information, significantly diminishes the reasoning abilities of models.
no code implementations • 9 Jan 2024 • Haoyi Xiong, Xuhong LI, Xiaofei Zhang, Jiamin Chen, Xinhao Sun, Yuchen Li, Zeyi Sun, Mengnan Du
Given the complexity and lack of transparency in deep neural networks (DNNs), extensive efforts have been made to make these systems more interpretable or explain their behaviors in accessible terms.
1 code implementation • 23 Dec 2023 • Guanchu Wang, Yu-Neng Chuang, Fan Yang, Mengnan Du, Chia-Yuan Chang, Shaochen Zhong, Zirui Liu, Zhaozhuo Xu, Kaixiong Zhou, Xuanting Cai, Xia Hu
This meta-attribution leverages the versatility of generic backbone encoders to comprehensively encode the attribution knowledge for the input instance, which enables TVE to seamlessly transfer to explain various downstream tasks, without the need for training on task-specific data.
no code implementations • 19 Oct 2023 • Hua Tang, Lu Cheng, Ninghao Liu, Mengnan Du
While the accuracy-fairness trade-off has been frequently observed in the literature of fair machine learning, rigorous theoretical analyses have been scarce.
no code implementations • 17 Sep 2023 • Zirui He, Huiqi Deng, Haiyan Zhao, Ninghao Liu, Mengnan Du
Recent research has shown that large language models rely on spurious correlations in the data for natural language understanding (NLU) tasks.
Natural Language Understanding Out-of-Distribution Generalization
1 code implementation • 15 Sep 2023 • Zhihao Hu, Yiran Xu, Mengnan Du, Jindong Gu, Xinmei Tian, Fengxiang He
Our adaptive reweighing method prioritizes samples closer to the decision boundary and assigns a higher weight to improve the generalizability of fair classifiers.
no code implementations • 2 Sep 2023 • Haiyan Zhao, Hanjie Chen, Fan Yang, Ninghao Liu, Huiqi Deng, Hengyi Cai, Shuaiqiang Wang, Dawei Yin, Mengnan Du
For each paradigm, we summarize the goals and dominant approaches for generating local explanations of individual predictions and global explanations of overall model knowledge.
no code implementations • 20 Aug 2023 • Yingji Li, Mengnan Du, Rui Song, Xin Wang, Ying Wang
Large Language Models (LLMs) have shown powerful performance and development prospects and are widely deployed in the real world.
1 code implementation • 8 Aug 2023 • Zihan Guan, Mengnan Du, Ninghao Liu
An emerging detection strategy in the vision and NLP domains is based on an intriguing phenomenon: when training models on a mixture of backdoor and clean samples, the loss on backdoor samples drops significantly faster than on clean samples, allowing backdoor samples to be easily detected by selecting samples with the lowest loss values.
no code implementations • 14 Jul 2023 • Chia-Yuan Chang, Yu-Neng Chuang, Guanchu Wang, Mengnan Du, Na Zou
Domain generalization aims to learn a generalization model that can perform well on unseen test domains by only training on limited source domains.
no code implementations • 4 Jul 2023 • Yingji Li, Mengnan Du, Xin Wang, Ying Wang
Meanwhile, experimental results on the GLUE benchmark show that CCPA retains the language modeling capability of PLMs.
no code implementations • 27 Jun 2023 • Tianlin Li, Qing Guo, Aishan Liu, Mengnan Du, Zhiming Li, Yang Liu
Existing fairness regularization terms fail to achieve decision rationale alignment because they only constrain last-layer outputs while ignoring intermediate neuron alignment.
1 code implementation • ICLR 2022 • Qizhang Feng, Ninghao Liu, Fan Yang, Ruixiang Tang, Mengnan Du, Xia Hu
Graph Neural Networks (GNNs) are gaining extensive attention for their application in graph data.
1 code implementation • NeurIPS 2023 • Yucheng Shi, Mengnan Du, Xuansheng Wu, Zihan Guan, Jin Sun, Ninghao Liu
Defending against such attacks is challenging, especially for real-world black-box models where only query access is permitted.
no code implementations • 13 Mar 2023 • Chenyang Li, Jihoon Chung, Mengnan Du, Haimin Wang, Xianlian Zhou, Bo Shen
This paper focuses on two model compression techniques: low-rank approximation and weight pruning in neural networks, which are very popular nowadays.
no code implementations • 2 Mar 2023 • Huiqi Deng, Na Zou, Mengnan Du, Weifu Chen, Guocan Feng, Ziwei Yang, Zheyang Li, Quanshi Zhang
Various attribution methods have been developed to explain deep neural networks (DNNs) by inferring the attribution/importance/contribution score of each input variable to the final output.
no code implementations • 7 Feb 2023 • Yu-Neng Chuang, Guanchu Wang, Fan Yang, Zirui Liu, Xuanting Cai, Mengnan Du, Xia Hu
Finally, we summarize the challenges of deploying XAI acceleration methods to real-world scenarios, overcoming the trade-off between faithfulness and efficiency, and the selection of different acceleration methods.
Explainable artificial intelligence Explainable Artificial Intelligence (XAI)
no code implementations • 26 Nov 2022 • Yu-Neng Chuang, Kwei-Herng Lai, Ruixiang Tang, Mengnan Du, Chia-Yuan Chang, Na Zou, Xia Hu
Knowledge graph data are prevalent in real-world applications, and knowledge graph neural networks (KGNNs) are essential techniques for knowledge graph representation learning.
no code implementations • 25 Aug 2022 • Mengnan Du, Fengxiang He, Na Zou, DaCheng Tao, Xia Hu
We first introduce the concepts of shortcut learning of language models.
1 code implementation • 20 Jul 2022 • Guanchu Wang, Mengnan Du, Ninghao Liu, Na Zou, Xia Hu
Existing work on fairness modeling commonly assumes that sensitive attributes for all instances are fully available, which may not be true in many real-world applications due to the high cost of acquiring sensitive information.
no code implementations • 29 Jun 2022 • Qizhang Feng, Mengnan Du, Na Zou, Xia Hu
The digitization of healthcare data coupled with advances in computational capabilities has propelled the adoption of machine learning (ML) in healthcare.
1 code implementation • 17 Jun 2022 • Guanchu Wang, Yu-Neng Chuang, Mengnan Du, Fan Yang, Quan Zhou, Pushkar Tripathi, Xuanting Cai, Xia Hu
Even though Shapley value provides an effective explanation for a DNN model prediction, the computation relies on the enumeration of all possible input feature coalitions, which leads to the exponentially growing complexity.
no code implementations • 19 Jan 2022 • Zhiming Li, Yanzhou Li, Tianlin Li, Mengnan Du, Bozhi Wu, Yushi Cao, Junzhe Jiang, Yang Liu
We propose a Cond-Idf measurement to interpret this behavior, which quantifies the relatedness of a token with a label and its project-specificness.
no code implementations • 16 Oct 2021 • Mengnan Du, Subhabrata Mukherjee, Yu Cheng, Milad Shokouhi, Xia Hu, Ahmed Hassan Awadallah
Recent work has focused on compressing pre-trained language models (PLMs) like BERT where the major focus has been to improve the in-distribution performance for downstream tasks.
no code implementations • 29 Sep 2021 • Ruixiang Tang, Hongye Jin, Curtis Wigington, Mengnan Du, Rajiv Jain, Xia Hu
The main idea is to insert a watermark which is only known to defender into the protected model and the watermark will then be transferred into all stolen models.
no code implementations • NeurIPS 2021 • Mengnan Du, Subhabrata Mukherjee, Guanchu Wang, Ruixiang Tang, Ahmed Hassan Awadallah, Xia Hu
This process not only requires a lot of instance-level annotations for sensitive attributes, it also does not guarantee that all fairness sensitive information has been removed from the encoder.
no code implementations • 28 May 2021 • Huiqi Deng, Na Zou, Mengnan Du, Weifu Chen, Guocan Feng, Xia Hu
However, the attribution problem has not been well-defined, which lacks a unified guideline to the contribution assignment process.
no code implementations • 17 May 2021 • Yuening Li, Zhengzhang Chen, Daochen Zha, Mengnan Du, Denghui Zhang, Haifeng Chen, Xia Hu
Motivated by the success of disentangled representation learning in computer vision, we study the possibility of learning semantic-rich time-series representations, which remains unexplored due to three main challenges: 1) sequential data structure introduces complex temporal correlations and makes the latent representations hard to interpret, 2) sequential models suffer from KL vanishing problem, and 3) interpretable semantic concepts for time-series often rely on multiple factors instead of individuals.
no code implementations • 14 Apr 2021 • Huiqi Deng, Na Zou, Weifu Chen, Guocan Feng, Mengnan Du, Xia Hu
The basic idea is to learn a source signal by back-propagation such that the mutual information between input and output should be as much as possible preserved in the mutual information between input and the source signal.
no code implementations • NAACL 2021 • Mengnan Du, Varun Manjunatha, Rajiv Jain, Ruchi Deshpande, Franck Dernoncourt, Jiuxiang Gu, Tong Sun, Xia Hu
These two observations are further employed to formulate a measurement which can quantify the shortcut degree of each training sample.
no code implementations • 18 Jan 2021 • Fan Yang, Ninghao Liu, Mengnan Du, Xia Hu
With the wide use of deep neural networks (DNN), model interpretability has become a critical concern, since explainable decisions are preferred in high-stake scenarios.
no code implementations • 17 Nov 2020 • Ruixiang Tang, Mengnan Du, Xia Hu
In this paper, we present DSN (Deep Serial Number), a simple yet effective watermarking algorithm designed specifically for deep neural networks (DNNs).
no code implementations • 21 Aug 2020 • Huiqi Deng, Na Zou, Mengnan Du, Weifu Chen, Guocan Feng, Xia Hu
Attribution methods have been developed to understand the decision-making process of machine learning models, especially deep neural networks, by assigning importance scores to individual features.
no code implementations • 24 Jul 2020 • Sina Mohseni, Fan Yang, Shiva Pentyala, Mengnan Du, Yi Liu, Nic Lupfer, Xia Hu, Shuiwang Ji, Eric Ragan
Combating fake news and misinformation propagation is a challenging task in the post-truth era.
1 code implementation • 15 Jun 2020 • Ruixiang Tang, Mengnan Du, Yuening Li, Zirui Liu, Na Zou, Xia Hu
Image captioning has made substantial progress with huge supporting image collections sourced from the web.
1 code implementation • 15 Jun 2020 • Ruixiang Tang, Mengnan Du, Ninghao Liu, Fan Yang, Xia Hu
In this paper, we investigate a specific security problem called trojan attack, which aims to attack deployed DNN systems relying on the hidden trigger patterns inserted by malicious hackers.
no code implementations • 23 Apr 2020 • Ninghao Liu, Mengnan Du, Ruocheng Guo, Huan Liu, Xia Hu
In this paper, we review recent work on adversarial attacks and defenses, particularly from the perspective of machine learning interpretation.
9 code implementations • 3 Oct 2019 • Haofan Wang, Zifan Wang, Mengnan Du, Fan Yang, Zijian Zhang, Sirui Ding, Piotr Mardziel, Xia Hu
Recently, increasing attention has been drawn to the internal mechanisms of convolutional neural networks, and the reason why the network makes specific decisions.
Ranked #2 on Error Understanding on CUB-200-2011 (ResNet-101)
1 code implementation • 1 Oct 2019 • Yijun Bian, Qingquan Song, Mengnan Du, Jun Yao, Huanhuan Chen, Xia Hu
Neural architecture search (NAS) is gaining more and more attention in recent years due to its flexibility and remarkable capability to reduce the burden of neural network design.
no code implementations • 25 Sep 2019 • Weijie Fu, Meng Wang, Mengnan Du, Ninghao Liu, Shijie Hao, Xia Hu
Existing local explanation methods provide an explanation for each decision of black-box classifiers, in the form of relevance scores of features according to their contributions.
no code implementations • 13 Sep 2019 • Mengnan Du, Shiva Pentyala, Yuening Li, Xia Hu
The analysis further shows that LAE outperforms the state-of-the-arts by 6. 52%, 12. 03%, and 3. 08% respectively on three deepfake detection tasks in terms of generalization accuracy on previously unseen manipulations.
no code implementations • 23 Aug 2019 • Mengnan Du, Fan Yang, Na Zou, Xia Hu
Deep learning is increasingly being used in high-stake decision making applications that affect individual lives.
no code implementations • 13 Aug 2019 • Mengnan Du, Ninghao Liu, Fan Yang, Xia Hu
Recent explainability related studies have shown that state-of-the-art DNNs do not always adopt correct evidences to make decisions.
no code implementations • 11 Aug 2019 • Yuening Li, Xiao Huang, Jundong Li, Mengnan Du, Na Zou
SpecAE leverages Laplacian sharpening to amplify the distances between representations of anomalies and the ones of the majority.
no code implementations • 11 Aug 2019 • Yuening Li, Ninghao Liu, Jundong Li, Mengnan Du, Xia Hu
To this end, we propose a novel deep structured anomaly detection framework to identify the cross-modal anomalies embedded in the data.
no code implementations • 16 Jul 2019 • Fan Yang, Mengnan Du, Xia Hu
Interpretable Machine Learning (IML) has become increasingly important in many real-world applications, such as autonomous cars and medical diagnosis, where explanations are significantly preferred to help people better understand how machine learning systems work and further enhance their trust towards systems.
BIG-bench Machine Learning Interpretable Machine Learning +1
no code implementations • 8 Jul 2019 • Fan Yang, Shiva K. Pentyala, Sina Mohseni, Mengnan Du, Hao Yuan, Rhema Linder, Eric D. Ragan, Shuiwang Ji, Xia Hu
In this demo paper, we present the XFake system, an explainable fake news detector that assists end-users to identify news credibility.
no code implementations • 27 Mar 2019 • Mengnan Du, Ninghao Liu, Fan Yang, Shuiwang Ji, Xia Hu
REAT decomposes the final prediction of a RNN into additive contribution of each word in the input text.
no code implementations • 31 Jul 2018 • Mengnan Du, Ninghao Liu, Xia Hu
Interpretable machine learning tackles the important problem that humans cannot understand the behaviors of complex machine learning models and how these models arrive at a particular decision.
no code implementations • 19 Mar 2018 • Mengnan Du, Ninghao Liu, Qingquan Song, Xia Hu
While deep neural networks (DNN) have become an effective computational tool, the prediction results are often criticized by the lack of interpretability, which is essential in many real-world applications such as health informatics.