Search Results for author: Jindong Gu

Found 98 papers, 41 papers with code

Does Machine Unlearning Truly Remove Model Knowledge? A Framework for Auditing Unlearning in LLMs

no code implementations29 May 2025 Haokun Chen, Yueqi Zhang, Yuan Bi, Yao Zhang, Tong Liu, Jinhe Bi, Jian Lan, Jindong Gu, Claudia Grosser, Denis Krompass, Nassir Navab, Volker Tresp

In this work, we introduce a comprehensive auditing framework for unlearning evaluation, comprising three benchmark datasets, six unlearning algorithms, and five prompt-based auditing methods.

Machine Unlearning

A Comprehensive Survey in LLM(-Agent) Full Stack Safety: Data, Training and Deployment

no code implementations22 Apr 2025 Kun Wang, Guibin Zhang, Zhenhong Zhou, Jiahao Wu, Miao Yu, Shiqian Zhao, Chenlong Yin, Jinhu Fu, Yibo Yan, Hanjun Luo, Liang Lin, Zhihao Xu, Haolang Lu, Xinye Cao, Xinyun Zhou, Weifei Jin, Fanci Meng, Junyuan Mao, Yu Wang, Hao Wu, Minghe Wang, Fan Zhang, Junfeng Fang, Wenjie Qu, Yue Liu, Chengwei Liu, Yifan Zhang, Qiankun Li, Chongye Guo, Yalan Qin, Zhaoxin Fan, Yi Ding, Donghai Hong, Jiaming Ji, Yingxin Lai, Zitong Yu, Xinfeng Li, Yifan Jiang, Yanhui Li, Xinyu Deng, Junlin Wu, Dongxia Wang, Yihao Huang, Yufei Guo, Jen-tse Huang, Qiufeng Wang, Wenxuan Wang, Dongrui Liu, Yanwei Yue, Wenke Huang, Guancheng Wan, Heng Chang, Tianlin Li, Yi Yu, Chenghao Li, Jiawei Li, Lei Bai, Jie Zhang, Qing Guo, Jingyi Wang, Tianlong Chen, Joey Tianyi Zhou, Xiaojun Jia, Weisong Sun, Cong Wu, Jing Chen, Xuming Hu, Yiming Li, Xiao Wang, Ningyu Zhang, Luu Anh Tuan, Guowen Xu, Jiaheng Zhang, Tianwei Zhang, Xingjun Ma, Jindong Gu, Xiang Wang, Bo An, Jun Sun, Mohit Bansal, Shirui Pan, Lingjuan Lyu, Yuval Elovici, Bhavya Kailkhura, Yaodong Yang, Hongwei Li, Wenyuan Xu, Yizhou Sun, Wei Wang, Qing Li, Ke Tang, Yu-Gang Jiang, Felix Juefei-Xu, Hui Xiong, XiaoFeng Wang, DaCheng Tao, Philip S. Yu, Qingsong Wen, Yang Liu

Currently, existing surveys on LLM safety primarily focus on specific stages of the LLM lifecycle, e. g., deployment phase or fine-tuning phase, lacking a comprehensive understanding of the entire "lifechain" of LLMs.

Model Editing

Exploring Typographic Visual Prompts Injection Threats in Cross-Modality Generation Models

no code implementations14 Mar 2025 Hao Cheng, Erjia Xiao, Yichi Wang, Kaidi Xu, Mengshu Sun, Jindong Gu, Renjing Xu

To better observe performance modifications and characteristics of this threat, we also introduce the TVPI Dataset.

Magnet: Multi-turn Tool-use Data Synthesis and Distillation via Graph Translation

no code implementations10 Mar 2025 Fan Yin, Zifeng Wang, I-Hung Hsu, Jun Yan, Ke Jiang, Yanfei Chen, Jindong Gu, Long T. Le, Kai-Wei Chang, Chen-Yu Lee, Hamid Palangi, Tomas Pfister

To address this, we propose Magnet, a principled framework for synthesizing high-quality training trajectories to enhance the function calling capability of large language model agents in multi-turn conversations with humans.

Large Language Model

Improving Adversarial Transferability in MLLMs via Dynamic Vision-Language Alignment Attack

no code implementations27 Feb 2025 Chenhe Gu, Jindong Gu, Andong Hua, Yao Qin

Multimodal Large Language Models (MLLMs), built upon LLMs, have recently gained attention for their capabilities in image recognition and understanding.

PlanGEN: A Multi-Agent Framework for Generating Planning and Reasoning Trajectories for Complex Problem Solving

no code implementations22 Feb 2025 Mihir Parmar, Xin Liu, Palash Goyal, Yanfei Chen, Long Le, Swaroop Mishra, Hossein Mobahi, Jindong Gu, Zifeng Wang, Hootan Nakhost, Chitta Baral, Chen-Yu Lee, Tomas Pfister, Hamid Palangi

Recent agent frameworks and inference-time algorithms often struggle with complex planning problems due to limitations in verifying generated plans or reasoning and varying complexity of instances within a single task.

Jailbreak-AudioBench: In-Depth Evaluation and Analysis of Jailbreak Threats for Large Audio Language Models

no code implementations23 Jan 2025 Hao Cheng, Erjia Xiao, Jing Shao, Yichi Wang, Le Yang, Chao Shen, Philip Torr, Jindong Gu, Renjing Xu

Integrating various modality encoders further expands their capabilities, giving rise to Multimodal Large Language Models (MLLMs) that process not only text but also visual and auditory modality inputs.

Safety Alignment

FocalPO: Enhancing Preference Optimizing by Focusing on Correct Preference Rankings

no code implementations11 Jan 2025 Tong Liu, Xiao Yu, Wenxuan Zhou, Jindong Gu, Volker Tresp

These algorithms implicitly treat the LLM as a reward model, and focus on training it to correct misranked preference pairs.

Open Eyes, Then Reason: Fine-grained Visual Mathematical Understanding in MLLMs

1 code implementation11 Jan 2025 Shan Zhang, Aotian Chen, Yanpeng Sun, Jindong Gu, Yi-Yu Zheng, Piotr Koniusz, Kai Zou, Anton Van Den Hengel, Yuan Xue

Current multimodal large language models (MLLMs) often underperform on mathematical problem-solving tasks that require fine-grained visual understanding.

Math Mathematical Problem-Solving +2

SafetyDPO: Scalable Safety Alignment for Text-to-Image Generation

no code implementations13 Dec 2024 Runtao Liu, Chen I Chieh, Jindong Gu, Jipeng Zhang, Renjie Pi, Qifeng Chen, Philip Torr, Ashkan Khakzar, Fabio Pizzati

Using a custom DPO strategy and this dataset, we train safety experts, in the form of low-rank adaptation (LoRA) matrices, able to guide the generation process away from specific safety-related concepts.

Safety Alignment Text to Image Generation +1

Not Just Text: Uncovering Vision Modality Typographic Threats in Image Generation Models

no code implementations CVPR 2025 Hao Cheng, Erjia Xiao, Jiayan Yang, Jiahang Cao, Qiang Zhang, Jize Zhang, Kaidi Xu, Jindong Gu, Renjing Xu

Current image generation models can effortlessly produce high-quality, highly realistic images, but this also increases the risk of misuse.

Image Generation

FedBiP: Heterogeneous One-Shot Federated Learning with Personalized Latent Diffusion Models

no code implementations CVPR 2025 Haokun Chen, Hang Li, Yao Zhang, Jinhe Bi, Gengyuan Zhang, Yueqi Zhang, Philip Torr, Jindong Gu, Denis Krompass, Volker Tresp

However, directly applying pretrained LDM to heterogeneous OSFL results in significant distribution shifts in synthetic data, leading to performance degradation in classification models trained on such data.

Federated Learning

Visual Question Decomposition on Multimodal Large Language Models

no code implementations28 Sep 2024 Haowei Zhang, Jianzhe Liu, Zhen Han, Shuo Chen, Bailan He, Volker Tresp, Zhiqiang Xu, Jindong Gu

The finetuning pipeline consists of our proposed dataset and a training objective for selective decomposition.

Visual Question Answering (VQA)

Multimodal Pragmatic Jailbreak on Text-to-image Models

no code implementations27 Sep 2024 Tong Liu, Zhixin Lai, Gengyuan Zhang, Philip Torr, Vera Demberg, Volker Tresp, Jindong Gu

This work introduces a novel type of jailbreak, which triggers T2I models to generate the image with visual text, where the image and the text, although considered to be safe in isolation, combine to form unsafe content.

Manipulation Facing Threats: Evaluating Physical Vulnerabilities in End-to-End Vision Language Action Models

no code implementations20 Sep 2024 Hao Cheng, Erjia Xiao, Chengyuan Yu, Zhao Yao, Jiahang Cao, Qiang Zhang, Jiaxu Wang, Mengshu Sun, Kaidi Xu, Jindong Gu, Renjing Xu

Recently, driven by advancements in Multimodal Large Language Models (MLLMs), Vision Language Action Models (VLAMs) are being proposed to achieve better performance in open-vocabulary scenarios for robotic manipulation tasks.

Vision-Language-Action

HTS-Attack: Heuristic Token Search for Jailbreaking Text-to-Image Models

no code implementations25 Aug 2024 Sensen Gao, Xiaojun Jia, Yihao Huang, Ranjie Duan, Jindong Gu, Yang Bai, Yang Liu, Qing Guo

Text-to-Image(T2I) models have achieved remarkable success in image generation and editing, yet these models still have many potential issues, particularly in generating inappropriate or Not-Safe-For-Work(NSFW) content.

Heuristic Search Image Generation +2

Can Editing LLMs Inject Harm?

1 code implementation29 Jul 2024 Canyu Chen, Baixiang Huang, Zekun Li, Zhaorun Chen, Shiyang Lai, Xiongxiao Xu, Jia-Chen Gu, Jindong Gu, Huaxiu Yao, Chaowei Xiao, Xifeng Yan, William Yang Wang, Philip Torr, Dawn Song, Kai Shu

Then, we find that editing attacks can inject both types of misinformation into LLMs, and the effectiveness is particularly high for commonsense misinformation injection.

Fairness General Knowledge +4

Dataset Distillation by Automatic Training Trajectories

1 code implementation19 Jul 2024 Dai Liu, Jindong Gu, Hu Cao, Carsten Trinitis, Martin Schulz

Dataset Distillation is used to create a concise, yet informative, synthetic dataset that can replace the original dataset for training purposes.

Dataset Distillation

MMRo: Are Multimodal LLMs Eligible as the Brain for In-Home Robotics?

no code implementations28 Jun 2024 Jinming Li, Yichen Zhu, Zhiyuan Xu, Jindong Gu, Minjie Zhu, Xin Liu, Ning Liu, Yaxin Peng, Feifei Feng, Jian Tang

It is fundamentally challenging for robots to serve as useful assistants in human environments because this requires addressing a spectrum of sub-problems across robotics, including perception, language understanding, reasoning, and planning.

Task Planning Visual Reasoning

Localizing Events in Videos with Multimodal Queries

no code implementations CVPR 2025 Gengyuan Zhang, Mang Ling Ada Fok, Jialu Ma, Yan Xia, Daniel Cremers, Philip Torr, Volker Tresp, Jindong Gu

Localizing events in videos based on semantic queries is a pivotal task in video understanding, with the growing significance of user-oriented applications like video search.

Natural Language Queries Video Understanding

Provably Better Explanations with Optimized Aggregation of Feature Attributions

no code implementations7 Jun 2024 Thomas Decker, Ananta R. Bhattarai, Jindong Gu, Volker Tresp, Florian Buettner

Using feature attributions for post-hoc explanations is a common practice to understand and verify the predictions of opaque machine learning models.

Learning Visual Prompts for Guiding the Attention of Vision Transformers

no code implementations5 Jun 2024 Razieh Rezaei, Masoud Jalili Sabet, Jindong Gu, Daniel Rueckert, Philip Torr, Ashkan Khakzar

The learned visual prompt, added to any input image would redirect the attention of the pre-trained vision transformer to its spatial location on the image.

Visual Prompting

Improved Techniques for Optimization-Based Jailbreaking on Large Language Models

1 code implementation31 May 2024 Xiaojun Jia, Tianyu Pang, Chao Du, Yihao Huang, Jindong Gu, Yang Liu, Xiaochun Cao, Min Lin

Many red-teaming efforts aim to jailbreak LLMs, where among these efforts, the Greedy Coordinate Gradient (GCG) attack's success has led to a growing interest in the study of optimization-based jailbreaking techniques.

Red Teaming

Typography Leads Semantic Diversifying: Amplifying Adversarial Transferability across Multimodal Large Language Models

no code implementations30 May 2024 Hao Cheng, Erjia Xiao, Jiayan Yang, Jiahang Cao, Qiang Zhang, Le Yang, Jize Zhang, Kaidi Xu, Jindong Gu, Renjing Xu

Recently, Multimodal Large Language Models (MLLMs) achieve remarkable performance in numerous zero-shot tasks due to their outstanding cross-modal interaction and comprehension abilities.

Diversity

Special Characters Attack: Toward Scalable Training Data Extraction From Large Language Models

no code implementations9 May 2024 Yang Bai, Ge Pei, Jindong Gu, Yong Yang, Xingjun Ma

In this paper, we take a step further and show that certain special characters or their combinations with English letters are stronger memory triggers, leading to more severe data leakage.

Energy-Latency Manipulation of Multi-modal Large Language Models via Verbose Samples

no code implementations25 Apr 2024 Kuofeng Gao, Jindong Gu, Yang Bai, Shu-Tao Xia, Philip Torr, Wei Liu, Zhifeng Li

For verbose videos, a frame feature diversity loss is proposed to increase the feature diversity among frames.

Diversity

A Survey on Responsible Generative AI: What to Generate and What Not

no code implementations8 Apr 2024 Jindong Gu

To answer the question, this paper investigates the practical responsible requirements of both textual and visual generative models, outlining five key considerations: generating truthful content, avoiding toxic content, refusing harmful instruction, leaking no training data-related content, and ensuring generated content identifiable.

Red Teaming GPT-4V: Are GPT-4V Safe Against Uni/Multi-Modal Jailbreak Attacks?

1 code implementation4 Apr 2024 Shuo Chen, Zhen Han, Bailan He, Zifeng Ding, Wenqian Yu, Philip Torr, Volker Tresp, Jindong Gu

Various jailbreak attacks have been proposed to red-team Large Language Models (LLMs) and revealed the vulnerable safeguards of LLMs.

Red Teaming

Which Model Generated This Image? A Model-Agnostic Approach for Origin Attribution

1 code implementation3 Apr 2024 Fengyuan Liu, Haochen Luo, Yiming Li, Philip Torr, Jindong Gu

In this work, we study the origin attribution of generated images in a practical setting where only a few images generated by a source model are available and the source model cannot be accessed.

model One-Class Classification

An Image Is Worth 1000 Lies: Adversarial Transferability across Prompts on Vision-Language Models

1 code implementation14 Mar 2024 Haochen Luo, Jindong Gu, Fengyuan Liu, Philip Torr

Given that VLMs rely on prompts to adapt to different tasks, an intriguing question emerges: Can a single adversarial image mislead all predictions of VLMs when a thousand different prompts are given?

Hide in Thicket: Generating Imperceptible and Rational Adversarial Perturbations on 3D Point Clouds

1 code implementation CVPR 2024 Tianrui Lou, Xiaojun Jia, Jindong Gu, Li Liu, Siyuan Liang, Bangyan He, Xiaochun Cao

We find that concealing deformation perturbations in areas insensitive to human eyes can achieve a better trade-off between imperceptibility and adversarial strength, specifically in parts of the object surface that are complex and exhibit drastic curvature changes.

3D Point Cloud Classification Adversarial Attack +1

Unveiling Typographic Deceptions: Insights of the Typographic Vulnerability in Large Vision-Language Model

no code implementations29 Feb 2024 Hao Cheng, Erjia Xiao, Jindong Gu, Le Yang, Jinhao Duan, Jize Zhang, Jiahang Cao, Kaidi Xu, Renjing Xu

Large Vision-Language Models (LVLMs) rely on vision encoders and Large Language Models (LLMs) to exhibit remarkable capabilities on various multi-modal tasks in the joint space of vision and language.

Language Modeling Language Modelling +2

Stop Reasoning! When Multimodal LLM with Chain-of-Thought Reasoning Meets Adversarial Image

1 code implementation22 Feb 2024 Zefeng Wang, Zhen Han, Shuo Chen, Fan Xue, Zifeng Ding, Xun Xiao, Volker Tresp, Philip Torr, Jindong Gu

Based on our findings, we further propose a novel attack method, termed as stop-reasoning attack, that attacks the model while bypassing the CoT reasoning process.

Adversarial Robustness Multimodal Reasoning +1

Can Large Language Model Agents Simulate Human Trust Behavior?

1 code implementation7 Feb 2024 Chengxing Xie, Canyu Chen, Feiran Jia, Ziyu Ye, Shiyang Lai, Kai Shu, Jindong Gu, Adel Bibi, Ziniu Hu, David Jurgens, James Evans, Philip Torr, Bernard Ghanem, Guohao Li

In this paper, we focus on one critical and elemental behavior in human interactions, trust, and investigate whether LLM agents can simulate human trust behavior.

Language Modeling Language Modelling +1

Inducing High Energy-Latency of Large Vision-Language Models with Verbose Images

1 code implementation20 Jan 2024 Kuofeng Gao, Yang Bai, Jindong Gu, Shu-Tao Xia, Philip Torr, Zhifeng Li, Wei Liu

Once attackers maliciously induce high energy consumption and latency time (energy-latency cost) during inference of VLMs, it will exhaust computational resources.

Diversity

Does Few-shot Learning Suffer from Backdoor Attacks?

no code implementations31 Dec 2023 Xinwei Liu, Xiaojun Jia, Jindong Gu, Yuan Xun, Siyuan Liang, Xiaochun Cao

However, in this paper, we propose the Few-shot Learning Backdoor Attack (FLBA) to show that FSL can still be vulnerable to backdoor attacks.

Backdoor Attack Few-Shot Learning

XAI for In-hospital Mortality Prediction via Multimodal ICU Data

1 code implementation29 Dec 2023 Xingqiao Li, Jindong Gu, Zhiyong Wang, Yancheng Yuan, Bo Du, Fengxiang He

To address this issue, this paper proposes an eXplainable Multimodal Mortality Predictor (X-MMP) approaching an efficient, explainable AI solution for predicting in-hospital mortality via multimodal ICU data.

Decision Making Mortality Prediction

Initialization Matters for Adversarial Transfer Learning

1 code implementation CVPR 2024 Andong Hua, Jindong Gu, Zhiyu Xue, Nicholas Carlini, Eric Wong, Yao Qin

Based on this, we propose Robust Linear Initialization (RoLI) for adversarial finetuning, which initializes the linear head with the weights obtained by adversarial linear probing to maximally inherit the robustness from pretraining.

Adversarial Robustness image-classification +2

OT-Attack: Enhancing Adversarial Transferability of Vision-Language Models via Optimal Transport Optimization

no code implementations7 Dec 2023 Dongchen Han, Xiaojun Jia, Yang Bai, Jindong Gu, Yang Liu, Xiaochun Cao

Investigating the generation of high-transferability adversarial examples is crucial for uncovering VLP models' vulnerabilities in practical scenarios.

Adversarial Attack Data Augmentation +2

TranSegPGD: Improving Transferability of Adversarial Examples on Semantic Segmentation

no code implementations3 Dec 2023 Xiaojun Jia, Jindong Gu, Yihao Huang, Simeng Qin, Qing Guo, Yang Liu, Xiaochun Cao

At the second stage, the pixels are divided into different branches based on their transferable property which is dependent on Kullback-Leibler divergence.

Adversarial Attack image-classification +3

Improving Adversarial Transferability via Model Alignment

1 code implementation30 Nov 2023 Avery Ma, Amir-Massoud Farahmand, Yangchen Pan, Philip Torr, Jindong Gu

During the alignment process, the parameters of the source model are fine-tuned to minimize an alignment loss.

model

Can Multimodal Large Language Models Truly Perform Multimodal In-Context Learning?

no code implementations29 Nov 2023 Shuo Chen, Zhen Han, Bailan He, Jianzhe Liu, Mark Buckley, Yao Qin, Philip Torr, Volker Tresp, Jindong Gu

Experiments revealed that multimodal ICL is predominantly driven by the textual content whereas the visual information in the demos has little influence.

In-Context Learning

MM-SafetyBench: A Benchmark for Safety Evaluation of Multimodal Large Language Models

2 code implementations29 Nov 2023 Xin Liu, Yichen Zhu, Jindong Gu, Yunshi Lan, Chao Yang, Yu Qiao

The security concerns surrounding Large Language Models (LLMs) have been extensively explored, yet the safety of Multimodal Large Language Models (MLLMs) remains understudied.

Benchmarking Robustness of Text-Image Composed Retrieval

1 code implementation24 Nov 2023 Shitong Sun, Jindong Gu, Shaogang Gong

In this paper, we perform the first robustness study and establish three new diversified benchmarks for systematic analysis of text-image composed retrieval against natural corruptions in both vision and text and further probe textural understanding.

Attribute Benchmarking +2

SPOT! Revisiting Video-Language Models for Event Understanding

no code implementations21 Nov 2023 Gengyuan Zhang, Jinhe Bi, Jindong Gu, Yanyu Chen, Volker Tresp

This raises a question: with such weak supervision, can video representation in video-language models gain the ability to distinguish even factual discrepancies in textual description and understand fine-grained events?

Attribute Video Understanding

Fast Propagation is Better: Accelerating Single-Step Adversarial Training via Sampling Subnetworks

1 code implementation24 Oct 2023 Xiaojun Jia, Jianshu Li, Jindong Gu, Yang Bai, Xiaochun Cao

Besides, we provide theoretical analysis to show the model robustness can be improved by the single-step adversarial training with sampled subnetworks.

Boosting Fair Classifier Generalization through Adaptive Priority Reweighing

1 code implementation15 Sep 2023 Zhihao Hu, Yiran Xu, Mengnan Du, Jindong Gu, Xinmei Tian, Fengxiang He

Our adaptive reweighing method prioritizes samples closer to the decision boundary and assigns a higher weight to improve the generalizability of fair classifiers.

Decision Making Fairness

Exploring Non-additive Randomness on ViT against Query-Based Black-Box Attacks

no code implementations12 Sep 2023 Jindong Gu, Fangyun Wei, Philip Torr, Han Hu

In this work, we first taxonomize the stochastic defense strategies against QBBA.

Multi-event Video-Text Retrieval

1 code implementation ICCV 2023 Gengyuan Zhang, Jisen Ren, Jindong Gu, Volker Tresp

In this study, we introduce the Multi-event Video-Text Retrieval (MeVTR) task, addressing scenarios in which each video contains multiple different events, as a niche scenario of the conventional Video-Text Retrieval Task.

Language Modelling Text Retrieval +1

FedDAT: An Approach for Foundation Model Finetuning in Multi-Modal Heterogeneous Federated Learning

1 code implementation21 Aug 2023 Haokun Chen, Yao Zhang, Denis Krompass, Jindong Gu, Volker Tresp

FedDAT is the first approach that enables an efficient distributed finetuning of foundation models for a variety of heterogeneous Vision-Language tasks.

Federated Learning Knowledge Distillation +1

FedPop: Federated Population-based Hyperparameter Tuning

2 code implementations16 Aug 2023 Haokun Chen, Denis Krompass, Jindong Gu, Volker Tresp

This is mainly because their "training-after-tuning" framework is unsuitable for FL with limited client computation power.

Computational Efficiency Evolutionary Algorithms +1

A Systematic Survey of Prompt Engineering on Vision-Language Foundation Models

2 code implementations24 Jul 2023 Jindong Gu, Zhen Han, Shuo Chen, Ahmad Beirami, Bailan He, Gengyuan Zhang, Ruotong Liao, Yao Qin, Volker Tresp, Philip Torr

This paper aims to provide a comprehensive survey of cutting-edge research in prompt engineering on three types of vision-language models: multimodal-to-text generation models (e. g. Flamingo), image-text matching models (e. g.

Image-text matching Language Modeling +6

Reliable Evaluation of Adversarial Transferability

no code implementations14 Jun 2023 Wenqian Yu, Jindong Gu, Zhijiang Li, Philip Torr

Adversarial examples (AEs) with small adversarial perturbations can mislead deep neural networks (DNNs) into wrong predictions.

Towards Robust Prompts on Vision-Language Models

no code implementations17 Apr 2023 Jindong Gu, Ahmad Beirami, Xuezhi Wang, Alex Beutel, Philip Torr, Yao Qin

With the advent of vision-language models (VLMs) that can perform in-context and prompt-based learning, how can we design prompting approaches that robustly generalize to distribution shift and can be used on novel classes outside the support set of the prompts?

In-Context Learning Prompt Learning

Backdoor Defense via Adaptively Splitting Poisoned Dataset

1 code implementation CVPR 2023 Kuofeng Gao, Yang Bai, Jindong Gu, Yong Yang, Shu-Tao Xia

With the split clean data pool and polluted data pool, ASD successfully defends against backdoor attacks during training.

backdoor defense

Influencer Backdoor Attack on Semantic Segmentation

1 code implementation21 Mar 2023 Haoheng Lan, Jindong Gu, Philip Torr, Hengshuang Zhao

In this work, we explore backdoor attacks on segmentation models to misclassify all pixels of a victim class by injecting a specific trigger on non-victim pixels during inferences, which is dubbed Influencer Backdoor Attack (IBA).

Backdoor Attack Position +2

Explainability and Robustness of Deep Visual Classification Models

no code implementations3 Jan 2023 Jindong Gu

The vulnerability of deep neural networks poses challenges to current visual classification models.

Classification image-classification +2

Do DALL-E and Flamingo Understand Each Other?

no code implementations ICCV 2023 Hang Li, Jindong Gu, Rajat Koner, Sahand Sharifzadeh, Volker Tresp

To study this question, we propose a reconstruction task where Flamingo generates a description for a given image and DALL-E uses this description as input to synthesize a new image.

Image Captioning Image Reconstruction +5

SegPGD: An Effective and Efficient Adversarial Attack for Evaluating and Boosting Segmentation Robustness

2 code implementations25 Jul 2022 Jindong Gu, Hengshuang Zhao, Volker Tresp, Philip Torr

Since SegPGD can create more effective adversarial examples, the adversarial training with our SegPGD can boost the robustness of segmentation models.

Adversarial Attack Segmentation +1

Towards Efficient Adversarial Training on Vision Transformers

no code implementations21 Jul 2022 Boxi Wu, Jindong Gu, Zhifeng Li, Deng Cai, Xiaofei He, Wei Liu

Vision Transformer (ViT), as a powerful alternative to Convolutional Neural Network (CNN), has received much attention.

Watermark Vaccine: Adversarial Attacks to Prevent Watermark Removal

1 code implementation17 Jul 2022 Xinwei Liu, Jian Liu, Yang Bai, Jindong Gu, Tao Chen, Xiaojun Jia, Xiaochun Cao

Inspired by the vulnerability of DNNs on adversarial perturbations, we propose a novel defence mechanism by adversarial machine learning for good.

FRAug: Tackling Federated Learning with Non-IID Features via Representation Augmentation

1 code implementation ICCV 2023 Haokun Chen, Ahmed Frikha, Denis Krompass, Jindong Gu, Volker Tresp

Real-world applications usually involve a distribution shift across the datasets of the different clients, which hurts the generalization ability of the clients to unseen samples from their respective data distributions.

Federated Learning

ECOLA: Enhanced Temporal Knowledge Embeddings with Contextualized Language Representations

no code implementations17 Mar 2022 Zhen Han, Ruotong Liao, Jindong Gu, Yao Zhang, Zifeng Ding, Yujia Gu, Heinz Köppl, Hinrich Schütze, Volker Tresp

Since conventional knowledge embedding models cannot take full advantage of the abundant textual information, there have been extensive research efforts in enhancing knowledge embedding using texts.

Knowledge Graph Embedding Link Prediction +1

Adversarial Examples on Segmentation Models Can be Easy to Transfer

no code implementations22 Nov 2021 Jindong Gu, Hengshuang Zhao, Volker Tresp, Philip Torr

The high transferability achieved by our method shows that, in contrast to the observations in previous work, adversarial examples on a segmentation model can be easy to transfer to other segmentation models.

Adversarial Robustness Attribute +6

Are Vision Transformers Robust to Patch Perturbations?

no code implementations20 Nov 2021 Jindong Gu, Volker Tresp, Yao Qin

However, when ViTs are attacked by an adversary, the attention mechanism can be easily fooled to focus more on the adversarially perturbed patches and cause a mistake.

image-classification Image Classification

Are Vision Transformers Robust to Patch-wise Perturbations?

no code implementations29 Sep 2021 Jindong Gu, Volker Tresp, Yao Qin

Based on extensive qualitative and quantitative experiments, we discover that ViT's stronger robustness to natural corrupted patches and higher vulnerability against adversarial patches are both caused by the attention mechanism.

image-classification Image Classification

Simple Distillation Baselines for Improving Small Self-supervised Models

1 code implementation21 Jun 2021 Jindong Gu, Wei Liu, Yonglong Tian

While large self-supervised models have rivalled the performance of their supervised counterparts, small models still struggle.

Attacking Adversarial Attacks as A Defense

no code implementations9 Jun 2021 Boxi Wu, Heng Pan, Li Shen, Jindong Gu, Shuai Zhao, Zhifeng Li, Deng Cai, Xiaofei He, Wei Liu

In this work, we find that the adversarial attacks can also be vulnerable to small perturbations.

Quantifying Predictive Uncertainty in Medical Image Analysis with Deep Kernel Learning

1 code implementation1 Jun 2021 Zhiliang Wu, Yinchong Yang, Jindong Gu, Volker Tresp

We propose an uncertainty-aware deep kernel learning model which permits the estimation of the uncertainty in the prediction by a pipeline of a Convolutional Neural Network and a sparse Gaussian Process.

Medical Image Analysis Prediction

Capsule Network is Not More Robust than Convolutional Network

no code implementations CVPR 2021 Jindong Gu, Volker Tresp, Han Hu

The examination reveals five major new/different components in CapsNet: a transformation process, a dynamic routing layer, a squashing function, a marginal loss other than cross-entropy loss, and an additional class-conditional reconstruction loss for regularization.

image-classification Image Classification

Effective and Efficient Vote Attack on Capsule Networks

1 code implementation ICLR 2021 Jindong Gu, Baoyuan Wu, Volker Tresp

As alternatives to CNNs, the recently proposed Capsule Networks (CapsNets) are shown to be more robust to white-box attacks than CNNs under popular attack protocols.

Adversarial Robustness

Interpretable Graph Capsule Networks for Object Recognition

no code implementations3 Dec 2020 Jindong Gu, Volker Tresp

In the proposed model, individual classification explanations can be created effectively and efficiently.

Adversarial Robustness Object +1

Introspective Learning by Distilling Knowledge from Online Self-explanation

no code implementations19 Sep 2020 Jindong Gu, Zhiliang Wu, Volker Tresp

Motivated by the conclusion, we propose an implementation of introspective learning by distilling knowledge from online self-explanations.

Knowledge Distillation

Search for Better Students to Learn Distilled Knowledge

no code implementations30 Jan 2020 Jindong Gu, Volker Tresp

The knowledge of a well-performed teacher is distilled to a student with a small architecture.

Knowledge Distillation Model Compression

Neural Network Memorization Dissection

no code implementations21 Nov 2019 Jindong Gu, Volker Tresp

What is the difference between DNNs trained with random labels and the ones trained with true labels?

Memorization

Improving the Robustness of Capsule Networks to Image Affine Transformations

no code implementations CVPR 2020 Jindong Gu, Volker Tresp

Our investigation reveals that the routing procedure contributes neither to the generalization ability nor to the affine robustness of the CapsNets.

Semantics for Global and Local Interpretation of Deep Neural Networks

no code implementations21 Oct 2019 Jindong Gu, Volker Tresp

Deep neural networks (DNNs) with high expressiveness have achieved state-of-the-art performance in many tasks.

Understanding Bias in Machine Learning

no code implementations2 Sep 2019 Jindong Gu, Daniela Oelke

Bias is known to be an impediment to fair decisions in many domains such as human resources, the public sector, health care etc.

BIG-bench Machine Learning

Saliency Methods for Explaining Adversarial Attacks

no code implementations22 Aug 2019 Jindong Gu, Volker Tresp

The idea behind saliency methods is to explain the classification decisions of neural networks by creating so-called saliency maps.

General Classification

Understanding Individual Decisions of CNNs via Contrastive Backpropagation

2 code implementations5 Dec 2018 Jindong Gu, Yinchong Yang, Volker Tresp

The experiments and analysis conclude that the explanations generated by LRP are not class-discriminative.

General Classification

Semi-supervised Outlier Detection using Generative And Adversary Framework

no code implementations ICLR 2018 Jindong Gu, Matthias Schubert, Volker Tresp

In the adversarial process of training CorGAN, the Generator is supposed to generate outlier samples for negative class, and the Discriminator as an one-class classifier is trained to distinguish data from training datasets (i. e. positive class) and generated data from the Generator (i. e. negative class).

General Classification Multi-class Classification +2

Cannot find the paper you are looking for? You can Submit a new open access paper.