Search Results for author: Yugeng Liu

Found 3 papers, 2 papers with code

Comprehensive Assessment of Jailbreak Attacks Against LLMs

no code implementations8 Feb 2024 Junjie Chu, Yugeng Liu, Ziqing Yang, Xinyue Shen, Michael Backes, Yang Zhang

Some jailbreak prompt datasets, available from the Internet, can also achieve high attack success rates on many LLMs, such as ChatGLM3, GPT-3. 5, and PaLM2.

Ethics

Backdoor Attacks Against Dataset Distillation

2 code implementations3 Jan 2023 Yugeng Liu, Zheng Li, Michael Backes, Yun Shen, Yang Zhang

A model trained on this smaller distilled dataset can attain comparable performance to a model trained on the original training dataset.

Backdoor Attack

ML-Doctor: Holistic Risk Assessment of Inference Attacks Against Machine Learning Models

1 code implementation4 Feb 2021 Yugeng Liu, Rui Wen, Xinlei He, Ahmed Salem, Zhikun Zhang, Michael Backes, Emiliano De Cristofaro, Mario Fritz, Yang Zhang

As a result, we lack a comprehensive picture of the risks caused by the attacks, e. g., the different scenarios they can be applied to, the common factors that influence their performance, the relationship among them, or the effectiveness of possible defenses.

Attribute BIG-bench Machine Learning +3

Cannot find the paper you are looking for? You can Submit a new open access paper.