Search Results for author: Mengnan Du

Found 60 papers, 14 papers with code

Towards Explanation of DNN-based Prediction with Guided Feature Inversion

no code implementations19 Mar 2018 Mengnan Du, Ninghao Liu, Qingquan Song, Xia Hu

While deep neural networks (DNN) have become an effective computational tool, the prediction results are often criticized by the lack of interpretability, which is essential in many real-world applications such as health informatics.

Decision Making

Techniques for Interpretable Machine Learning

no code implementations31 Jul 2018 Mengnan Du, Ninghao Liu, Xia Hu

Interpretable machine learning tackles the important problem that humans cannot understand the behaviors of complex machine learning models and how these models arrive at a particular decision.

BIG-bench Machine Learning Interpretable Machine Learning

On Attribution of Recurrent Neural Network Predictions via Additive Decomposition

no code implementations27 Mar 2019 Mengnan Du, Ninghao Liu, Fan Yang, Shuiwang Ji, Xia Hu

REAT decomposes the final prediction of a RNN into additive contribution of each word in the input text.

Decision Making

XFake: Explainable Fake News Detector with Visualizations

no code implementations8 Jul 2019 Fan Yang, Shiva K. Pentyala, Sina Mohseni, Mengnan Du, Hao Yuan, Rhema Linder, Eric D. Ragan, Shuiwang Ji, Xia Hu

In this demo paper, we present the XFake system, an explainable fake news detector that assists end-users to identify news credibility.

Attribute

Evaluating Explanation Without Ground Truth in Interpretable Machine Learning

no code implementations16 Jul 2019 Fan Yang, Mengnan Du, Xia Hu

Interpretable Machine Learning (IML) has become increasingly important in many real-world applications, such as autonomous cars and medical diagnosis, where explanations are significantly preferred to help people better understand how machine learning systems work and further enhance their trust towards systems.

BIG-bench Machine Learning Interpretable Machine Learning +1

SpecAE: Spectral AutoEncoder for Anomaly Detection in Attributed Networks

no code implementations11 Aug 2019 Yuening Li, Xiao Huang, Jundong Li, Mengnan Du, Na Zou

SpecAE leverages Laplacian sharpening to amplify the distances between representations of anomalies and the ones of the majority.

Anomaly Detection Density Estimation

Deep Structured Cross-Modal Anomaly Detection

no code implementations11 Aug 2019 Yuening Li, Ninghao Liu, Jundong Li, Mengnan Du, Xia Hu

To this end, we propose a novel deep structured anomaly detection framework to identify the cross-modal anomalies embedded in the data.

Anomaly Detection

Learning Credible Deep Neural Networks with Rationale Regularization

no code implementations13 Aug 2019 Mengnan Du, Ninghao Liu, Fan Yang, Xia Hu

Recent explainability related studies have shown that state-of-the-art DNNs do not always adopt correct evidences to make decisions.

text-classification Text Classification

Fairness in Deep Learning: A Computational Perspective

no code implementations23 Aug 2019 Mengnan Du, Fan Yang, Na Zou, Xia Hu

Deep learning is increasingly being used in high-stake decision making applications that affect individual lives.

Decision Making Fairness

Towards Generalizable Deepfake Detection with Locality-aware AutoEncoder

no code implementations13 Sep 2019 Mengnan Du, Shiva Pentyala, Yuening Li, Xia Hu

The analysis further shows that LAE outperforms the state-of-the-arts by 6. 52%, 12. 03%, and 3. 08% respectively on three deepfake detection tasks in terms of generalization accuracy on previously unseen manipulations.

Active Learning DeepFake Detection +2

Distribution-Guided Local Explanation for Black-Box Classifiers

no code implementations25 Sep 2019 Weijie Fu, Meng Wang, Mengnan Du, Ninghao Liu, Shijie Hao, Xia Hu

Existing local explanation methods provide an explanation for each decision of black-box classifiers, in the form of relevance scores of features according to their contributions.

Sub-Architecture Ensemble Pruning in Neural Architecture Search

1 code implementation1 Oct 2019 Yijun Bian, Qingquan Song, Mengnan Du, Jun Yao, Huanhuan Chen, Xia Hu

Neural architecture search (NAS) is gaining more and more attention in recent years due to its flexibility and remarkable capability to reduce the burden of neural network design.

Ensemble Learning Ensemble Pruning +1

Score-CAM: Score-Weighted Visual Explanations for Convolutional Neural Networks

9 code implementations3 Oct 2019 Haofan Wang, Zifan Wang, Mengnan Du, Fan Yang, Zijian Zhang, Sirui Ding, Piotr Mardziel, Xia Hu

Recently, increasing attention has been drawn to the internal mechanisms of convolutional neural networks, and the reason why the network makes specific decisions.

Adversarial Attack Decision Making +1

Adversarial Attacks and Defenses: An Interpretation Perspective

no code implementations23 Apr 2020 Ninghao Liu, Mengnan Du, Ruocheng Guo, Huan Liu, Xia Hu

In this paper, we review recent work on adversarial attacks and defenses, particularly from the perspective of machine learning interpretation.

Adversarial Attack Adversarial Defense +2

Mitigating Gender Bias in Captioning Systems

1 code implementation15 Jun 2020 Ruixiang Tang, Mengnan Du, Yuening Li, Zirui Liu, Na Zou, Xia Hu

Image captioning has made substantial progress with huge supporting image collections sourced from the web.

Benchmarking Gender Prediction +1

An Embarrassingly Simple Approach for Trojan Attack in Deep Neural Networks

1 code implementation15 Jun 2020 Ruixiang Tang, Mengnan Du, Ninghao Liu, Fan Yang, Xia Hu

In this paper, we investigate a specific security problem called trojan attack, which aims to attack deployed DNN systems relying on the hidden trigger patterns inserted by malicious hackers.

A Unified Taylor Framework for Revisiting Attribution Methods

no code implementations21 Aug 2020 Huiqi Deng, Na Zou, Mengnan Du, Weifu Chen, Guocan Feng, Xia Hu

Attribution methods have been developed to understand the decision-making process of machine learning models, especially deep neural networks, by assigning importance scores to individual features.

Benchmarking Decision Making

Deep Serial Number: Computational Watermarking for DNN Intellectual Property Protection

no code implementations17 Nov 2020 Ruixiang Tang, Mengnan Du, Xia Hu

In this paper, we present DSN (Deep Serial Number), a simple yet effective watermarking algorithm designed specifically for deep neural networks (DNNs).

Knowledge Distillation valid

Generative Counterfactuals for Neural Networks via Attribute-Informed Perturbation

no code implementations18 Jan 2021 Fan Yang, Ninghao Liu, Mengnan Du, Xia Hu

With the wide use of deep neural networks (DNN), model interpretability has become a critical concern, since explainable decisions are preferred in high-stake scenarios.

Attribute

Towards Interpreting and Mitigating Shortcut Learning Behavior of NLU Models

no code implementations NAACL 2021 Mengnan Du, Varun Manjunatha, Rajiv Jain, Ruchi Deshpande, Franck Dernoncourt, Jiuxiang Gu, Tong Sun, Xia Hu

These two observations are further employed to formulate a measurement which can quantify the shortcut degree of each training sample.

Mutual Information Preserving Back-propagation: Learn to Invert for Faithful Attribution

no code implementations14 Apr 2021 Huiqi Deng, Na Zou, Weifu Chen, Guocan Feng, Mengnan Du, Xia Hu

The basic idea is to learn a source signal by back-propagation such that the mutual information between input and output should be as much as possible preserved in the mutual information between input and the source signal.

Decision Making

Learning Disentangled Representations for Time Series

no code implementations17 May 2021 Yuening Li, Zhengzhang Chen, Daochen Zha, Mengnan Du, Denghui Zhang, Haifeng Chen, Xia Hu

Motivated by the success of disentangled representation learning in computer vision, we study the possibility of learning semantic-rich time-series representations, which remains unexplored due to three main challenges: 1) sequential data structure introduces complex temporal correlations and makes the latent representations hard to interpret, 2) sequential models suffer from KL vanishing problem, and 3) interpretable semantic concepts for time-series often rely on multiple factors instead of individuals.

Disentanglement Time Series +1

A General Taylor Framework for Unifying and Revisiting Attribution Methods

no code implementations28 May 2021 Huiqi Deng, Na Zou, Mengnan Du, Weifu Chen, Guocan Feng, Xia Hu

However, the attribution problem has not been well-defined, which lacks a unified guideline to the contribution assignment process.

Benchmarking Decision Making

Fairness via Representation Neutralization

no code implementations NeurIPS 2021 Mengnan Du, Subhabrata Mukherjee, Guanchu Wang, Ruixiang Tang, Ahmed Hassan Awadallah, Xia Hu

This process not only requires a lot of instance-level annotations for sensitive attributes, it also does not guarantee that all fairness sensitive information has been removed from the encoder.

Attribute Classification +1

Was my Model Stolen? Feature Sharing for Robust and Transferable Watermarks

no code implementations29 Sep 2021 Ruixiang Tang, Hongye Jin, Curtis Wigington, Mengnan Du, Rajiv Jain, Xia Hu

The main idea is to insert a watermark which is only known to defender into the protected model and the watermark will then be transferred into all stolen models.

Model extraction

Robustness Challenges in Model Distillation and Pruning for Natural Language Understanding

no code implementations16 Oct 2021 Mengnan Du, Subhabrata Mukherjee, Yu Cheng, Milad Shokouhi, Xia Hu, Ahmed Hassan Awadallah

Recent work has focused on compressing pre-trained language models (PLMs) like BERT where the major focus has been to improve the in-distribution performance for downstream tasks.

Knowledge Distillation Model Compression +1

Unveiling Project-Specific Bias in Neural Code Models

no code implementations19 Jan 2022 Zhiming Li, Yanzhou Li, Tianlin Li, Mengnan Du, Bozhi Wu, Yushi Cao, Junzhe Jiang, Yang Liu

We propose a Cond-Idf measurement to interpret this behavior, which quantifies the relatedness of a token with a label and its project-specificness.

Adversarial Robustness Vulnerability Detection

Accelerating Shapley Explanation via Contributive Cooperator Selection

1 code implementation17 Jun 2022 Guanchu Wang, Yu-Neng Chuang, Mengnan Du, Fan Yang, Quan Zhou, Pushkar Tripathi, Xuanting Cai, Xia Hu

Even though Shapley value provides an effective explanation for a DNN model prediction, the computation relies on the enumeration of all possible input feature coalitions, which leads to the exponentially growing complexity.

Fair Machine Learning in Healthcare: A Review

no code implementations29 Jun 2022 Qizhang Feng, Mengnan Du, Na Zou, Xia Hu

The digitization of healthcare data coupled with advances in computational capabilities has propelled the adoption of machine learning (ML) in healthcare.

BIG-bench Machine Learning Fairness

Mitigating Algorithmic Bias with Limited Annotations

1 code implementation20 Jul 2022 Guanchu Wang, Mengnan Du, Ninghao Liu, Na Zou, Xia Hu

Existing work on fairness modeling commonly assumes that sensitive attributes for all instances are fully available, which may not be true in many real-world applications due to the high cost of acquiring sensitive information.

Fairness

Mitigating Relational Bias on Knowledge Graphs

no code implementations26 Nov 2022 Yu-Neng Chuang, Kwei-Herng Lai, Ruixiang Tang, Mengnan Du, Chia-Yuan Chang, Na Zou, Xia Hu

Knowledge graph data are prevalent in real-world applications, and knowledge graph neural networks (KGNNs) are essential techniques for knowledge graph representation learning.

Graph Representation Learning Knowledge Graphs +1

Efficient XAI Techniques: A Taxonomic Survey

no code implementations7 Feb 2023 Yu-Neng Chuang, Guanchu Wang, Fan Yang, Zirui Liu, Xuanting Cai, Mengnan Du, Xia Hu

Finally, we summarize the challenges of deploying XAI acceleration methods to real-world scenarios, overcoming the trade-off between faithfulness and efficiency, and the selection of different acceleration methods.

Explainable artificial intelligence Explainable Artificial Intelligence (XAI)

Understanding and Unifying Fourteen Attribution Methods with Taylor Interactions

no code implementations2 Mar 2023 Huiqi Deng, Na Zou, Mengnan Du, Weifu Chen, Guocan Feng, Ziwei Yang, Zheyang Li, Quanshi Zhang

Various attribution methods have been developed to explain deep neural networks (DNNs) by inferring the attribution/importance/contribution score of each input variable to the final output.

FAIRER: Fairness as Decision Rationale Alignment

no code implementations27 Jun 2023 Tianlin Li, Qing Guo, Aishan Liu, Mengnan Du, Zhiming Li, Yang Liu

Existing fairness regularization terms fail to achieve decision rationale alignment because they only constrain last-layer outputs while ignoring intermediate neuron alignment.

Fairness

DISPEL: Domain Generalization via Domain-Specific Liberating

no code implementations14 Jul 2023 Chia-Yuan Chang, Yu-Neng Chuang, Guanchu Wang, Mengnan Du, Na Zou

Domain generalization aims to learn a generalization model that can perform well on unseen test domains by only training on limited source domains.

Domain Generalization

XGBD: Explanation-Guided Graph Backdoor Detection

1 code implementation8 Aug 2023 Zihan Guan, Mengnan Du, Ninghao Liu

An emerging detection strategy in the vision and NLP domains is based on an intriguing phenomenon: when training models on a mixture of backdoor and clean samples, the loss on backdoor samples drops significantly faster than on clean samples, allowing backdoor samples to be easily detected by selecting samples with the lowest loss values.

Attribute Graph Learning

A Survey on Fairness in Large Language Models

no code implementations20 Aug 2023 Yingji Li, Mengnan Du, Rui Song, Xin Wang, Ying Wang

Large Language Models (LLMs) have shown powerful performance and development prospects and are widely deployed in the real world.

Fairness

Explainability for Large Language Models: A Survey

no code implementations2 Sep 2023 Haiyan Zhao, Hanjie Chen, Fan Yang, Ninghao Liu, Huiqi Deng, Hengyi Cai, Shuaiqiang Wang, Dawei Yin, Mengnan Du

For each paradigm, we summarize the goals and dominant approaches for generating local explanations of individual predictions and global explanations of overall model knowledge.

Boosting Fair Classifier Generalization through Adaptive Priority Reweighing

1 code implementation15 Sep 2023 Zhihao Hu, Yiran Xu, Mengnan Du, Jindong Gu, Xinmei Tian, Fengxiang He

Our adaptive reweighing method prioritizes samples closer to the decision boundary and assigns a higher weight to improve the generalizability of fair classifiers.

Decision Making Fairness

Mitigating Shortcuts in Language Models with Soft Label Encoding

no code implementations17 Sep 2023 Zirui He, Huiqi Deng, Haiyan Zhao, Ninghao Liu, Mengnan Du

Recent research has shown that large language models rely on spurious correlations in the data for natural language understanding (NLU) tasks.

Natural Language Understanding Out-of-Distribution Generalization

A Theoretical Approach to Characterize the Accuracy-Fairness Trade-off Pareto Frontier

no code implementations19 Oct 2023 Hua Tang, Lu Cheng, Ninghao Liu, Mengnan Du

While the accuracy-fairness trade-off has been frequently observed in the literature of fair machine learning, rigorous theoretical analyses have been scarce.

Fairness

LETA: Learning Transferable Attribution for Generic Vision Explainer

no code implementations23 Dec 2023 Guanchu Wang, Yu-Neng Chuang, Fan Yang, Mengnan Du, Chia-Yuan Chang, Shaochen Zhong, Zirui Liu, Zhaozhuo Xu, Kaixiong Zhou, Xuanting Cai, Xia Hu

To address this problem, we develop a pre-trained, DNN-based, generic explainer on large-scale image datasets, and leverage its transferability to explain various vision models for downstream tasks.

Towards Explainable Artificial Intelligence (XAI): A Data Mining Perspective

no code implementations9 Jan 2024 Haoyi Xiong, Xuhong LI, Xiaofei Zhang, Jiamin Chen, Xinhao Sun, Yuchen Li, Zeyi Sun, Mengnan Du

Given the complexity and lack of transparency in deep neural networks (DNNs), extensive efforts have been made to make these systems more interpretable or explain their behaviors in accessible terms.

Data Valuation Decision Making +2

The Impact of Reasoning Step Length on Large Language Models

1 code implementation10 Jan 2024 Mingyu Jin, Qinkai Yu, Dong Shu, Haiyan Zhao, Wenyue Hua, Yanda Meng, Yongfeng Zhang, Mengnan Du

Alternatively, shortening the reasoning steps, even while preserving the key information, significantly diminishes the reasoning abilities of models.

Explanations of Classifiers Enhance Medical Image Segmentation via End-to-end Pre-training

no code implementations16 Jan 2024 Jiamin Chen, Xuhong LI, Yanwu Xu, Mengnan Du, Haoyi Xiong

Based on a large-scale medical image classification dataset, our work collects explanations from well-trained classifiers to generate pseudo labels of segmentation tasks.

Image Classification Image Segmentation +4

Explaining Time Series via Contrastive and Locally Sparse Perturbations

1 code implementation16 Jan 2024 Zichuan Liu, Yingying Zhang, Tianchun Wang, Zefan Wang, Dongsheng Luo, Mengnan Du, Min Wu, Yi Wang, Chunlin Chen, Lunting Fan, Qingsong Wen

Explaining multivariate time series is a compound challenge, as it requires identifying important locations in the time series and matching complex temporal patterns.

Contrastive Learning counterfactual +1

DataFrame QA: A Universal LLM Framework on DataFrame Question Answering Without Data Exposure

no code implementations27 Jan 2024 Junyi Ye, Mengnan Du, Guiling Wang

This paper introduces DataFrame question answering (QA), a novel task that utilizes large language models (LLMs) to generate Pandas queries for information retrieval and data analysis on dataframes, emphasizing safe and non-revealing data handling.

Information Retrieval Question Answering +1

Health-LLM: Personalized Retrieval-Augmented Disease Prediction System

1 code implementation1 Feb 2024 Mingyu Jin, Qinkai Yu, Dong Shu, Chong Zhang, Lizhou Fan, Wenyue Hua, Suiyuan Zhu, Yanda Meng, Zhenting Wang, Mengnan Du, Yongfeng Zhang

Compared to traditional health management applications, our system has three main advantages: (1) It integrates health reports and medical knowledge into a large model to ask relevant questions to large language model for disease prediction; (2) It leverages a retrieval augmented generation (RAG) mechanism to enhance feature extraction; (3) It incorporates a semi-automated feature updating framework that can merge and delete features to improve accuracy of disease prediction.

Disease Prediction Language Modelling +3

Large Language Models As Faithful Explainers

no code implementations7 Feb 2024 Yu-Neng Chuang, Guanchu Wang, Chia-Yuan Chang, Ruixiang Tang, Fan Yang, Mengnan Du, Xuanting Cai, Xia Hu

In this work, we introduce a generative explanation framework, xLLM, to improve the faithfulness of the explanations provided in natural language formats for LLMs.

Decision Making

Towards Uncovering How Large Language Model Works: An Explainability Perspective

no code implementations16 Feb 2024 Haiyan Zhao, Fan Yang, Bo Shen, Himabindu Lakkaraju, Mengnan Du

Large language models (LLMs) have led to breakthroughs in language tasks, yet the internal mechanisms that enable their remarkable generalization and reasoning abilities remain opaque.

Hallucination Language Modelling +3

What if LLMs Have Different World Views: Simulating Alien Civilizations with LLM-based Agents

no code implementations20 Feb 2024 Mingyu Jin, Beichen Wang, Zhaoqian Xue, Suiyuan Zhu, Wenyue Hua, Hua Tang, Kai Mei, Mengnan Du, Yongfeng Zhang

In this study, we introduce "CosmoAgent," an innovative artificial intelligence framework utilizing Large Language Models (LLMs) to simulate complex interactions between human and extraterrestrial civilizations, with a special emphasis on Stephen Hawking's cautionary advice about not sending radio signals haphazardly into the universe.

Decision Making

Knowledge Graph Large Language Model (KG-LLM) for Link Prediction

no code implementations12 Mar 2024 Dong Shu, Tianle Chen, Mingyu Jin, Yiting Zhang, Chong Zhang, Mengnan Du, Yongfeng Zhang

The task of predicting multiple links within knowledge graphs (KGs) stands as a challenge in the field of knowledge graph analysis, a challenge increasingly resolvable due to advancements in natural language processing (NLP) and KG embedding techniques.

In-Context Learning Knowledge Graphs +3

Usable XAI: 10 Strategies Towards Exploiting Explainability in the LLM Era

1 code implementation13 Mar 2024 Xuansheng Wu, Haiyan Zhao, Yaochen Zhu, Yucheng Shi, Fan Yang, Tianming Liu, Xiaoming Zhai, Wenlin Yao, Jundong Li, Mengnan Du, Ninghao Liu

Therefore, in this paper, we introduce Usable XAI in the context of LLMs by analyzing (1) how XAI can benefit LLMs and AI systems, and (2) how LLMs can contribute to the advancement of XAI.

Exploring Concept Depth: How Large Language Models Acquire Knowledge at Different Layers?

1 code implementation10 Apr 2024 Mingyu Jin, Qinkai Yu, Jingyuan Huang, Qingcheng Zeng, Zhenting Wang, Wenyue Hua, Haiyan Zhao, Kai Mei, Yanda Meng, Kaize Ding, Fan Yang, Mengnan Du, Yongfeng Zhang

We employ a probing technique to extract representations from different layers of the model and apply these to classification tasks.

Quantifying Multilingual Performance of Large Language Models Across Languages

no code implementations17 Apr 2024 Zihao Li, Yucheng Shi, Zirui Liu, Fan Yang, Ninghao Liu, Mengnan Du

However, currently there is no work to quantitatively measure the performance of LLMs in low-resource languages.

Cannot find the paper you are looking for? You can Submit a new open access paper.