Search Results for author: Xiaogeng Liu

Found 10 papers, 6 papers with code

JailBreakV-28K: A Benchmark for Assessing the Robustness of MultiModal Large Language Models against Jailbreak Attacks

no code implementations3 Apr 2024 Weidi Luo, Siyuan Ma, Xiaogeng Liu, XIAOYU GUO, Chaowei Xiao

With the rapid advancements in Multimodal Large Language Models (MLLMs), securing these models against malicious inputs while aligning them with human values has emerged as a critical challenge.

Don't Listen To Me: Understanding and Exploring Jailbreak Prompts of Large Language Models

no code implementations26 Mar 2024 Zhiyuan Yu, Xiaogeng Liu, Shunning Liang, Zach Cameron, Chaowei Xiao, Ning Zhang

Building on the insights from the user study, we also developed a system using AI as the assistant to automate the process of jailbreak prompt generation.

AdaShield: Safeguarding Multimodal Large Language Models from Structure-based Attack via Adaptive Shield Prompting

1 code implementation14 Mar 2024 Yu Wang, Xiaogeng Liu, Yu Li, Muhao Chen, Chaowei Xiao

However, with the integration of additional modalities, MLLMs are exposed to new vulnerabilities, rendering them prone to structured-based jailbreak attacks, where semantic content (e. g., "harmful text") has been injected into the images to mislead MLLMs.

Automatic and Universal Prompt Injection Attacks against Large Language Models

1 code implementation7 Mar 2024 Xiaogeng Liu, Zhiyuan Yu, Yizhe Zhang, Ning Zhang, Chaowei Xiao

Large Language Models (LLMs) excel in processing and generating human language, powered by their ability to interpret and follow instructions.

DeceptPrompt: Exploiting LLM-driven Code Generation via Adversarial Natural Language Instructions

no code implementations7 Dec 2023 Fangzhou Wu, Xiaogeng Liu, Chaowei Xiao

In this paper, we introduce DeceptPrompt, a novel algorithm that can generate adversarial natural language instructions that drive the Code LLMs to generate functionality correct code with vulnerabilities.

Code Generation

AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language Models

2 code implementations3 Oct 2023 Xiaogeng Liu, Nan Xu, Muhao Chen, Chaowei Xiao

In light of these challenges, we intend to answer this question: Can we develop an approach that can automatically generate stealthy jailbreak prompts?

Decision Making

Why Does Little Robustness Help? Understanding and Improving Adversarial Transferability from Surrogate Training

1 code implementation15 Jul 2023 Yechao Zhang, Shengshan Hu, Leo Yu Zhang, Junyu Shi, Minghui Li, Xiaogeng Liu, Wei Wan, Hai Jin

Building on these insights, we explore the impacts of data augmentation and gradient regularization on transferability and identify that the trade-off generally exists in the various training mechanisms, thus building a comprehensive blueprint for the regulation mechanism behind transferability.

Attribute Data Augmentation

Towards Efficient Data-Centric Robust Machine Learning with Noise-based Augmentation

no code implementations8 Mar 2022 Xiaogeng Liu, Haoyu Wang, Yechao Zhang, Fangzhou Wu, Shengshan Hu

The data-centric machine learning aims to find effective ways to build appropriate datasets which can improve the performance of AI models.

BIG-bench Machine Learning Data Augmentation

Protecting Facial Privacy: Generating Adversarial Identity Masks via Style-robust Makeup Transfer

1 code implementation CVPR 2022 Shengshan Hu, Xiaogeng Liu, Yechao Zhang, Minghui Li, Leo Yu Zhang, Hai Jin, Libing Wu

While deep face recognition (FR) systems have shown amazing performance in identification and verification, they also arouse privacy concerns for their excessive surveillance on users, especially for public face images widely spread on social networks.

Face Recognition

Cannot find the paper you are looking for? You can Submit a new open access paper.