1 code implementation • 23 Feb 2024 • Yihao Zhang, Hangzhou He, Jingyu Zhu, Huanran Chen, Yifei Wang, Zeming Wei
Instead of perturbing the samples, Sharpness-Aware Minimization (SAM) perturbs the model weights during training to find a more flat loss landscape and improve generalization.
1 code implementation • 9 Feb 2024 • Yichuan Mo, Yuji Wang, Zeming Wei, Yisen Wang
To our knowledge, we are the first to implement defense from the perspective of prompt tuning.
1 code implementation • 29 Dec 2023 • Julien Piet, Maha Alrashed, Chawin Sitawarin, Sizhe Chen, Zeming Wei, Elizabeth Sun, Basel Alomair, David Wagner
Jatmo only needs a task prompt and a dataset of inputs for the task: it uses the teacher model to generate outputs.
no code implementations • 10 Oct 2023 • Zeming Wei, Yifei Wang, Yisen Wang
Large Language Models (LLMs) have shown remarkable success in various tasks, but concerns about their safety and the potential for generating malicious content have emerged.
1 code implementation • 24 Jun 2023 • Zeming Wei, Xiyue Zhang, Yihao Zhang, Meng Sun
In this paper, we propose a novel framework of Weighted Finite Automata (WFA) extraction and explanation to tackle the limitations for natural language tasks.
1 code implementation • 9 May 2023 • Zeming Wei, Jingyu Zhu, Yihao Zhang
In this paper, we explore SAM in the context of adversarial robustness.
1 code implementation • 20 Apr 2023 • Yihao Zhang, Zeming Wei, Xiyue Zhang, Meng Sun
To evaluate the effectiveness of our implementation and improvements, we conduct extensive experiments on a set of benchmark datasets.
1 code implementation • CVPR 2023 • Zeming Wei, Yifei Wang, Yiwen Guo, Yisen Wang
Adversarial training has been widely acknowledged as the most effective method to improve the adversarial robustness against adversarial examples for Deep Neural Networks (DNNs).
1 code implementation • 27 Jun 2022 • Zeming Wei, Xiyue Zhang, Meng Sun
Compositional approaches that are scablable to natural languages fall short in extraction precision.