1 code implementation • 4 Dec 2024 • Haoyu Jiang, Xuhong Wang, Ping Yi, Shanzhe Lei, Yilun Lin
This paper proposes a multi-party credible watermarking framework (CredID) involving a trusted third party (TTP) and multiple LLM vendors to address these issues.
1 code implementation • 17 Nov 2024 • Haiyang Yu, Tian Xie, Jiaping Gui, Pengyang Wang, Ping Yi, Yue Wu
Given the diversity of modalities, BackdoorMBTI facilitates systematic evaluation across different data types.
no code implementations • 15 Jul 2024 • Xuhong Wang, Haoyu Jiang, Yi Yu, Jingru Yu, Yilun Lin, Ping Yi, Yingchun Wang, Yu Qiao, Li Li, Fei-Yue Wang
Large Language Models (LLMs) are increasingly integrated into diverse industries, posing substantial security risks due to unauthorized replication and misuse.
no code implementations • 28 May 2024 • Nan Li, Haoyu Jiang, Ping Yi
Deep Neural Networks (DNNs) are known to be vulnerable to backdoor attacks, posing concerning threats to their reliable deployment.
no code implementations • 28 May 2024 • Nan Li, Haiyang Yu, Ping Yi
Deep Neural Networks (DNNs) are known to be vulnerable to backdoor attacks, posing concerning threats to their reliable deployment.
no code implementations • 22 May 2024 • Pengzhou Cheng, Yidong Ding, Tianjie Ju, Zongru Wu, Wei Du, Ping Yi, Zhuosheng Zhang, Gongshen Liu
To improve the recall of the RAG for the target contexts, we introduce a knowledge graph to construct structured data to achieve hard matching at a fine-grained level.
1 code implementation • 4 Dec 2023 • Haoyu Jiang, Haiyang Yu, Nan Li, Ping Yi
We then pre-train a generative self-supervised graph autoencoder (GAE) to better learn the features of benign models in order to detect backdoor models without knowing the attack strategy.
no code implementations • 11 Mar 2021 • Haowen Liu, Ping Yi, Hsiao-Ying Lin, Jie Shi, Weidong Qiu
We propose DAFAR, a feedback framework that allows deep learning models to detect/purify adversarial examples in high effectiveness and universality, with low area and time overhead.