Search Results for author: Wenkai Yang

Found 14 papers, 10 papers with code

Exploring Backdoor Vulnerabilities of Chat Models

1 code implementation3 Apr 2024 Yunzhuo Hao, Wenkai Yang, Yankai Lin

Chat models are extensively adopted across various real-world scenarios, thus the security of chat models deserves increasing attention.

Backdoor Attack

Watch Out for Your Agents! Investigating Backdoor Threats to LLM-Based Agents

1 code implementation17 Feb 2024 Wenkai Yang, Xiaohan Bi, Yankai Lin, Sishuo Chen, Jie zhou, Xu sun

We first formulate a general framework of agent backdoor attacks, then we present a thorough analysis on the different forms of agent backdoor attacks.

Backdoor Attack Data Poisoning

Enabling Large Language Models to Learn from Rules

no code implementations15 Nov 2023 Wenkai Yang, Yankai Lin, Jie zhou, JiRong Wen

The current knowledge learning paradigm of LLMs is mainly based on learning from examples, in which LLMs learn the internal rule implicitly from a certain number of supervised examples.

Two Stream Scene Understanding on Graph Embedding

no code implementations12 Nov 2023 Wenkai Yang, Wenyuan Sun, Runxaing Huang

This architecture utilizes a graph feature stream and an image feature stream, aiming to merge the strengths of both modalities for improved performance in image classification and scene graph generation tasks.

Graph Attention Graph Embedding +4

Towards Codable Watermarking for Injecting Multi-bits Information to LLMs

1 code implementation29 Jul 2023 Lean Wang, Wenkai Yang, Deli Chen, Hao Zhou, Yankai Lin, Fandong Meng, Jie zhou, Xu sun

As large language models (LLMs) generate texts with increasing fluency and realism, there is a growing need to identify the source of texts to prevent the abuse of LLMs.

Language Modelling

Communication Efficient Federated Learning for Multilingual Neural Machine Translation with Adapter

1 code implementation21 May 2023 Yi Liu, Xiaohan Bi, Lei LI, Sishuo Chen, Wenkai Yang, Xu sun

However, as pre-trained language models (PLMs) continue to increase in size, the communication cost for transmitting parameters during synchronization has become a training speed bottleneck.

Clustering Federated Learning +2

Fine-Tuning Deteriorates General Textual Out-of-Distribution Detection by Distorting Task-Agnostic Features

2 code implementations30 Jan 2023 Sishuo Chen, Wenkai Yang, Xiaohan Bi, Xu sun

We find that: (1) no existing method behaves well in both settings; (2) fine-tuning PLMs on in-distribution data benefits detecting semantic shifts but severely deteriorates detecting non-semantic shifts, which can be attributed to the distortion of task-agnostic features.

Out-of-Distribution Detection Out of Distribution (OOD) Detection

Integrating Local Real Data with Global Gradient Prototypes for Classifier Re-Balancing in Federated Long-Tailed Learning

no code implementations25 Jan 2023 Wenkai Yang, Deli Chen, Hao Zhou, Fandong Meng, Jie zhou, Xu sun

Federated Learning (FL) has become a popular distributed learning paradigm that involves multiple clients training a global model collaboratively in a data privacy-preserving manner.

Federated Learning Privacy Preserving

When to Trust Aggregated Gradients: Addressing Negative Client Sampling in Federated Learning

no code implementations25 Jan 2023 Wenkai Yang, Yankai Lin, Guangxiang Zhao, Peng Li, Jie zhou, Xu sun

Federated Learning has become a widely-used framework which allows learning a global model on decentralized local datasets under the condition of protecting local data privacy.

Federated Learning text-classification +1

Expose Backdoors on the Way: A Feature-Based Efficient Defense against Textual Backdoor Attacks

1 code implementation14 Oct 2022 Sishuo Chen, Wenkai Yang, Zhiyuan Zhang, Xiaohan Bi, Xu sun

In this work, we take the first step to investigate the unconcealment of textual poisoned samples at the intermediate-feature level and propose a feature-based efficient online defense method.

backdoor defense Sentiment Analysis

RAP: Robustness-Aware Perturbations for Defending against Backdoor Attacks on NLP Models

1 code implementation EMNLP 2021 Wenkai Yang, Yankai Lin, Peng Li, Jie zhou, Xu sun

Motivated by this observation, we construct a word-based robustness-aware perturbation to distinguish poisoned samples from clean samples to defend against the backdoor attacks on natural language processing (NLP) models.

Sentiment Analysis

Well-classified Examples are Underestimated in Classification with Deep Neural Networks

1 code implementation13 Oct 2021 Guangxiang Zhao, Wenkai Yang, Xuancheng Ren, Lei LI, Yunfang Wu, Xu sun

The conventional wisdom behind learning deep classification models is to focus on bad-classified examples and ignore well-classified examples that are far from the decision boundary.

Graph Classification imbalanced classification +4

Rethinking Stealthiness of Backdoor Attack against NLP Models

1 code implementation ACL 2021 Wenkai Yang, Yankai Lin, Peng Li, Jie zhou, Xu sun

In this work, we point out a potential problem of current backdoor attacking research: its evaluation ignores the stealthiness of backdoor attacks, and most of existing backdoor attacking methods are not stealthy either to system deployers or to system users.

Backdoor Attack Data Augmentation +2

Be Careful about Poisoned Word Embeddings: Exploring the Vulnerability of the Embedding Layers in NLP Models

1 code implementation NAACL 2021 Wenkai Yang, Lei LI, Zhiyuan Zhang, Xuancheng Ren, Xu sun, Bin He

However, in this paper, we find that it is possible to hack the model in a data-free way by modifying one single word embedding vector, with almost no accuracy sacrificed on clean samples.

Backdoor Attack Data Poisoning +4

Cannot find the paper you are looking for? You can Submit a new open access paper.