Search Results for author: Wenqi Wei

Found 36 papers, 17 papers with code

Trustworthy Distributed AI Systems: Robustness, Privacy, and Governance

no code implementations2 Feb 2024 Wenqi Wei, Ling Liu

Emerging Distributed AI systems are revolutionizing big data computing and data processing capabilities with growing economic and societal impact.

Fairness

FinLLMs: A Framework for Financial Reasoning Dataset Generation with Large Language Models

no code implementations19 Jan 2024 Ziqiang Yuan, Kaiyuan Wang, Shoutai Zhu, Ye Yuan, Jingya Zhou, Yanlin Zhu, Wenqi Wei

To address the limited data resources and reduce the annotation cost, we introduce FinLLMs, a method for generating financial question-answering data based on common financial formulas using Large Language Models.

Question Answering

Imperio: Language-Guided Backdoor Attacks for Arbitrary Model Control

1 code implementation2 Jan 2024 Ka-Ho Chow, Wenqi Wei, Lei Yu

This paper proposes Imperio, which harnesses the language understanding capabilities of NLP models to enrich backdoor attacks.

Backdoor Attack Image Classification +1

Hierarchical Pruning of Deep Ensembles with Focal Diversity

1 code implementation17 Nov 2023 Yanzhao Wu, Ka-Ho Chow, Wenqi Wei, Ling Liu

This paper presents a novel deep ensemble pruning approach, which can efficiently identify smaller deep ensembles and provide higher ensemble accuracy than the entire deep ensemble of a large number of member networks.

Decision Making Ensemble Pruning

Exploring Model Learning Heterogeneity for Boosting Ensemble Robustness

1 code implementation3 Oct 2023 Yanzhao Wu, Ka-Ho Chow, Wenqi Wei, Ling Liu

We show that this two-tier heterogeneity driven ensemble construction method can compose an ensemble team that promotes high ensemble diversity and low negative correlation among member models of the ensemble, strengthening ensemble robustness against both negative examples and adversarial attacks.

object-detection Object Detection +1

Invisible Watermarking for Audio Generation Diffusion Models

2 code implementations22 Sep 2023 Xirong Cao, Xiang Li, Divyesh Jadav, Yanzhao Wu, Zhehui Chen, Chen Zeng, Wenqi Wei

Diffusion models have gained prominence in the image domain for their capabilities in data generation and transformation, achieving state-of-the-art performance in various tasks in both image and audio domains.

Audio Generation

Rethinking Learning Rate Tuning in the Era of Large Language Models

1 code implementation16 Sep 2023 Hongpeng Jin, Wenqi Wei, Xuyu Wang, Wenbin Zhang, Yanzhao Wu

Second, we present LRBench++ to benchmark learning rate policies and facilitate learning rate tuning for both traditional DNNs and LLMs.

Few-shot Multi-domain Knowledge Rearming for Context-aware Defence against Advanced Persistent Threats

no code implementations13 Jun 2023 Gaolei Li, YuanYuan Zhao, Wenqi Wei, Yuchen Liu

Secondly, to rearm current security strategies, an finetuning-based deployment mechanism is proposed to transfer learned knowledge into the student model, while minimizing the defense cost.

Meta-Learning Scheduling

Securing Distributed SGD against Gradient Leakage Threats

1 code implementation10 May 2023 Wenqi Wei, Ling Liu, Jingya Zhou, Ka-Ho Chow, Yanzhao Wu

Next, we present a gradient leakage resilient approach to securing distributed SGD in federated learning, with differential privacy controlled noise as the tool.

Federated Learning

STDLens: Model Hijacking-Resilient Federated Learning for Object Detection

1 code implementation CVPR 2023 Ka-Ho Chow, Ling Liu, Wenqi Wei, Fatih Ilhan, Yanzhao Wu

Based on the insights, we introduce a three-tier forensic framework to identify and expel Trojaned gradients and reclaim the performance over the course of FL.

Federated Learning object-detection +1

GNN-Ensemble: Towards Random Decision Graph Neural Networks

no code implementations20 Mar 2023 Wenqi Wei, Mu Qiao, Divyesh Jadav

In the meantime, we show that GNN-Ensemble can significantly improve the adversarial robustness against attacks on GNNs.

Adversarial Robustness Decision Making +1

Machine Learning for Synthetic Data Generation: A Review

no code implementations8 Feb 2023 Yingzhou Lu, Minjie Shen, Huazheng Wang, Xiao Wang, Capucine van Rechem, Wenqi Wei

In light of these challenges, the concept of synthetic data generation emerges as a promising alternative that allows for data sharing and utilization in ways that real-world data cannot facilitate.

Fairness Synthetic Data Generation

Adaptive Deep Neural Network Inference Optimization with EENet

1 code implementation15 Jan 2023 Fatih Ilhan, Ka-Ho Chow, Sihao Hu, Tiansheng Huang, Selim Tekin, Wenqi Wei, Yanzhao Wu, Myungjin Lee, Ramana Kompella, Hugo Latapie, Gaowen Liu, Ling Liu

Instead of having every sample go through all DNN layers during prediction, EENet learns an early exit scheduler, which can intelligently terminate the inference earlier for certain predictions, which the model has high confidence of early exit.

Inference Optimization Scheduling +1

Gradient Leakage Attack Resilient Deep Learning

no code implementations25 Dec 2021 Wenqi Wei, Ling Liu

Although deep learning with differential privacy is a defacto standard for publishing deep learning models with differential privacy guarantee, we show that differentially private algorithms with fixed privacy parameters are vulnerable against gradient leakage attacks.

Network Representation Learning: From Preprocessing, Feature Extraction to Node Embedding

no code implementations14 Oct 2021 Jingya Zhou, Ling Liu, Wenqi Wei, Jianxi Fan

This survey paper reviews the design principles and the different node embedding techniques for network representation learning over homogeneous networks.

Graph Mining Knowledge Graphs +3

Gradient-Leakage Resilient Federated Learning

1 code implementation2 Jul 2021 Wenqi Wei, Ling Liu, Yanzhao Wu, Gong Su, Arun Iyengar

This paper presents a gradient leakage resilient approach to privacy-preserving federated learning with per training example-based client differential privacy, coined as Fed-CDP.

Federated Learning Privacy Preserving

Boosting Ensemble Accuracy by Revisiting Ensemble Diversity Metrics

1 code implementation CVPR 2021 Yanzhao Wu, Ling Liu, Zhongwei Xie, Ka-Ho Chow, Wenqi Wei

Our new metrics significantly improve the intrinsic correlation between high ensemble diversity and high ensemble accuracy.

Ensemble Learning Ensemble Pruning +1

Robust Deep Learning Ensemble against Deception

no code implementations14 Sep 2020 Wenqi Wei, Ling Liu

Third, XEnsemble provides a suite of algorithms to combine input verification and output verification to protect the DNN prediction models from both adversarial examples and out of distribution inputs.

Adversarial Robustness Denoising +1

Bitcoin Transaction Forecasting with Deep Network Representation Learning

no code implementations15 Jul 2020 Wenqi Wei, Qi Zhang, Ling Liu

First, we explore three interesting properties between Bitcoin transaction accounts: topological connectivity pattern of Bitcoin accounts, transaction amount pattern, and transaction dynamics.

Representation Learning

Understanding Object Detection Through An Adversarial Lens

1 code implementation11 Jul 2020 Ka-Ho Chow, Ling Liu, Mehmet Emre Gursoy, Stacey Truex, Wenqi Wei, Yanzhao Wu

We demonstrate that the proposed framework can serve as a methodical benchmark for analyzing adversarial behaviors and risks in real-time object detection systems.

Adversarial Robustness Autonomous Vehicles +3

LDP-Fed: Federated Learning with Local Differential Privacy

no code implementations5 Jun 2020 Stacey Truex, Ling Liu, Ka-Ho Chow, Mehmet Emre Gursoy, Wenqi Wei

However, in federated learning model parameter updates are collected iteratively from each participant and consist of high dimensional, continuous values with high precision (10s of digits after the decimal point), making existing LDP protocols inapplicable.

Federated Learning

A Framework for Evaluating Gradient Leakage Attacks in Federated Learning

2 code implementations22 Apr 2020 Wenqi Wei, Ling Liu, Margaret Loper, Ka-Ho Chow, Mehmet Emre Gursoy, Stacey Truex, Yanzhao Wu

FL offers default client privacy by allowing clients to keep their sensitive data on local devices and to only share local training parameter updates with the federated server.

Federated Learning

TOG: Targeted Adversarial Objectness Gradient Attacks on Real-time Object Detection Systems

2 code implementations9 Apr 2020 Ka-Ho Chow, Ling Liu, Mehmet Emre Gursoy, Stacey Truex, Wenqi Wei, Yanzhao Wu

The rapid growth of real-time huge data capturing has pushed the deep learning and data analytic computing to the edge systems.

Autonomous Driving Object +4

Effects of Differential Privacy and Data Skewness on Membership Inference Vulnerability

no code implementations21 Nov 2019 Stacey Truex, Ling Liu, Mehmet Emre Gursoy, Wenqi Wei, Lei Yu

Second, through MPLens, we highlight how the vulnerability of pre-trained models under membership inference attack is not uniform across all classes, particularly when the training data itself is skewed.

Inference Attack Membership Inference Attack

Deep Neural Network Ensembles against Deception: Ensemble Diversity, Accuracy and Robustness

no code implementations29 Aug 2019 Ling Liu, Wenqi Wei, Ka-Ho Chow, Margaret Loper, Emre Gursoy, Stacey Truex, Yanzhao Wu

In this paper we first give an overview of the concept of ensemble diversity and examine the three types of ensemble diversity in the context of DNN classifiers.

Ensemble Learning

Denoising and Verification Cross-Layer Ensemble Against Black-box Adversarial Attacks

no code implementations21 Aug 2019 Ka-Ho Chow, Wenqi Wei, Yanzhao Wu, Ling Liu

Deep neural networks (DNNs) have demonstrated impressive performance on many challenging machine learning tasks.

Denoising

Secure and Utility-Aware Data Collection with Condensed Local Differential Privacy

no code implementations15 May 2019 Mehmet Emre Gursoy, Acar Tamersoy, Stacey Truex, Wenqi Wei, Ling Liu

In this paper, we address the small user population problem by introducing the concept of Condensed Local Differential Privacy (CLDP) as a specialization of LDP, and develop a suite of CLDP protocols that offer desirable statistical utility while preserving privacy.

Cryptography and Security Databases

A Comparative Measurement Study of Deep Learning as a Service Framework

1 code implementation29 Oct 2018 Yanzhao Wu, Ling Liu, Calton Pu, Wenqi Cao, Semih Sahin, Wenqi Wei, Qi Zhang

Big data powered Deep Learning (DL) and its applications have blossomed in recent years, fueled by three technological trends: a large amount of digitized data openly accessible, a growing number of DL software frameworks in open source and commercial markets, and a selection of affordable parallel computing hardware devices.

Adversarial Examples in Deep Learning: Characterization and Divergence

no code implementations29 Jun 2018 Wenqi Wei, Ling Liu, Margaret Loper, Stacey Truex, Lei Yu, Mehmet Emre Gursoy, Yanzhao Wu

The burgeoning success of deep learning has raised the security and privacy concerns as more and more tasks are accompanied with sensitive data.

Adversarial Attack

Towards Demystifying Membership Inference Attacks

1 code implementation28 Jun 2018 Stacey Truex, Ling Liu, Mehmet Emre Gursoy, Lei Yu, Wenqi Wei

Our empirical results additionally show that (1) using the type of target model under attack within the attack model may not increase attack effectiveness and (2) collaborative learning in federated systems exposes vulnerabilities to membership inference risks when the adversary is a participant in the federation.

Cryptography and Security

Cannot find the paper you are looking for? You can Submit a new open access paper.