Search Results for author: Zeyang Sha

Found 7 papers, 1 papers with code

Prompt Stealing Attacks Against Large Language Models

no code implementations20 Feb 2024 Zeyang Sha, Yang Zhang

Our proposed prompt stealing attack aims to steal these well-designed prompts based on the generated answers.

Prompt Engineering

Conversation Reconstruction Attack Against GPT Models

no code implementations5 Feb 2024 Junjie Chu, Zeyang Sha, Michael Backes, Yang Zhang

We then introduce two advanced attacks aimed at better reconstructing previous conversations, specifically the UNR attack and the PBU attack.

Reconstruction Attack Semantic Similarity +1

Comprehensive Assessment of Toxicity in ChatGPT

no code implementations3 Nov 2023 Boyang Zhang, Xinyue Shen, Wai Man Si, Zeyang Sha, Zeyuan Chen, Ahmed Salem, Yun Shen, Michael Backes, Yang Zhang

Moderating offensive, hateful, and toxic language has always been an important but challenging topic in the domain of safe use in NLP.

From Visual Prompt Learning to Zero-Shot Transfer: Mapping Is All You Need

no code implementations9 Mar 2023 Ziqing Yang, Zeyang Sha, Michael Backes, Yang Zhang

In this sense, we propose SeMap, a more effective mapping using the semantic alignment between the pre-trained model's knowledge and the downstream task.

Fine-Tuning Is All You Need to Mitigate Backdoor Attacks

no code implementations18 Dec 2022 Zeyang Sha, Xinlei He, Pascal Berrang, Mathias Humbert, Yang Zhang

Backdoor attacks represent one of the major threats to machine learning models.

DE-FAKE: Detection and Attribution of Fake Images Generated by Text-to-Image Generation Models

no code implementations13 Oct 2022 Zeyang Sha, Zheng Li, Ning Yu, Yang Zhang

To tackle this problem, we pioneer a systematic study on the detection and attribution of fake images generated by text-to-image generation models.

Attribute Fake Image Detection +1

Cannot find the paper you are looking for? You can Submit a new open access paper.