Search Results for author: Bang An

Found 29 papers, 16 papers with code

AegisLLM: Scaling Agentic Systems for Self-Reflective Defense in LLM Security

1 code implementation29 Apr 2025 Zikui Cai, Shayan Shabihi, Bang An, Zora Che, Brian R. Bartoldson, Bhavya Kailkhura, Tom Goldstein, Furong Huang

On the WMDP unlearning benchmark, AegisLLM achieves near-perfect unlearning with only 20 training examples and fewer than 300 LM calls.

PoisonedParrot: Subtle Data Poisoning Attacks to Elicit Copyright-Infringing Content from Large Language Models

no code implementations10 Mar 2025 Michael-Andrei Panaitescu-Liess, Pankayaraj Pathmanathan, Yigitcan Kaya, Zora Che, Bang An, Sicheng Zhu, Aakriti Agrawal, Furong Huang

In this paper, we introduce PoisonedParrot: the first stealthy data poisoning attack that induces an LLM to generate copyrighted content even when the model has not been directly trained on the specific copyrighted material.

Data Poisoning

LISA: Learning-Integrated Space Partitioning Framework for Traffic Accident Forecasting on Heterogeneous Spatiotemporal Data

1 code implementation19 Dec 2024 Bang An, Xun Zhou, Amin Vahedian, Nick Street, Jinping Guan, Jun Luo

Traffic accident forecasting is an important task for intelligent transportation management and emergency response systems.

GeoPro-Net: Learning Interpretable Spatiotemporal Prediction Models through Statistically-Guided Geo-Prototyping

1 code implementation19 Dec 2024 Bang An, Xun Zhou, Zirui Zhou, Ronilo Ragodos, Zenglin Xu, Jun Luo

Interpretation of the spatiotemporal forecasting mechanism is, however, challenging due to the complexity of multi-source spatiotemporal features, the non-intuitive nature of spatiotemporal patterns for non-expert users, and the presence of spatial heterogeneity in the data.

Second Language (Arabic) Acquisition of LLMs via Progressive Vocabulary Expansion

no code implementations16 Dec 2024 Jianqing Zhu, Huang Huang, Zhihang Lin, Juhao Liang, Zhengyang Tang, Khalid Almubarak, Abdulmohsen Alharthik, Bang An, Juncai He, Xiangbo Wu, Fei Yu, Junying Chen, Zhuoheng Ma, Yuhao Du, He Zhang, Emad A. Alghamdi, Lian Zhang, Ruoyu Sun, Haizhou Li, Benyou Wang, Jinchao Xu

This paper addresses the critical need for democratizing large language models (LLM) in the Arab world, a region that has seen slower progress in developing models comparable to state-of-the-art offerings like GPT-4 or ChatGPT 3. 5, due to a predominant focus on mainstream languages (e. g., English and Chinese).

Alignment at Pre-training! Towards Native Alignment for Arabic LLMs

1 code implementation4 Dec 2024 Juhao Liang, Zhenyang Cai, Jianqing Zhu, Huang Huang, Kewei Zong, Bang An, Mosen Alharthi, Juncai He, Lian Zhang, Haizhou Li, Benyou Wang, Jinchao Xu

The alignment of large language models (LLMs) is critical for developing effective and safe language models.

GenARM: Reward Guided Generation with Autoregressive Reward Model for Test-time Alignment

1 code implementation10 Oct 2024 Yuancheng Xu, Udari Madhushani Sehwag, Alec Koppel, Sicheng Zhu, Bang An, Furong Huang, Sumitra Ganesh

Traditional training-time methods finetune LLMs using human preference datasets but incur significant training costs and require repeated training to handle diverse user preferences.

Text Generation

SAFLEX: Self-Adaptive Augmentation via Feature Label Extrapolation

no code implementations3 Oct 2024 Mucong Ding, Bang An, Yuancheng Xu, Anirudh Satheesh, Furong Huang

Data augmentation, a cornerstone technique in deep learning, is crucial in enhancing model performance, especially with scarce labeled data.

Bilevel Optimization Data Augmentation +2

Can Watermarking Large Language Models Prevent Copyrighted Text Generation and Hide Training Data?

no code implementations24 Jul 2024 Michael-Andrei Panaitescu-Liess, Zora Che, Bang An, Yuancheng Xu, Pankayaraj Pathmanathan, Souradip Chakraborty, Sicheng Zhu, Tom Goldstein, Furong Huang

Surprisingly, we find that watermarking adversely affects the success rate of MIAs, complicating the task of detecting copyrighted text in the pretraining dataset.

Text Generation

Referee-Meta-Learning for Fast Adaptation of Locational Fairness

no code implementations20 Feb 2024 Weiye Chen, Yiqun Xie, Xiaowei Jia, Erhu He, Han Bao, Bang An, Xun Zhou

When dealing with data from distinct locations, machine learning algorithms tend to demonstrate an implicit preference of some locations over the others, which constitutes biases that sabotage the spatial fairness of the algorithm.

Decision Making Fairness +1

WAVES: Benchmarking the Robustness of Image Watermarks

1 code implementation16 Jan 2024 Bang An, Mucong Ding, Tahseen Rabbani, Aakriti Agrawal, Yuancheng Xu, ChengHao Deng, Sicheng Zhu, Abdirisak Mohamed, Yuxin Wen, Tom Goldstein, Furong Huang

Our evaluation examines two pivotal dimensions: the degree of image quality degradation and the efficacy of watermark detection after attacks.

Benchmarking

Explore Spurious Correlations at the Concept Level in Language Models for Text Classification

1 code implementation15 Nov 2023 YuHang Zhou, Paiheng Xu, Xiaoyu Liu, Bang An, Wei Ai, Furong Huang

We find that LMs, when encountering spurious correlations between a concept and a label in training or prompts, resort to shortcuts for predictions.

counterfactual In-Context Learning +2

C-Disentanglement: Discovering Causally-Independent Generative Factors under an Inductive Bias of Confounder

1 code implementation NeurIPS 2023 Xiaoyu Liu, Jiaxin Yuan, Bang An, Yuancheng Xu, Yifan Yang, Furong Huang

Representation learning assumes that real-world data is generated by a few semantically meaningful generative factors (i. e., sources of variation) and aims to discover them in the latent space.

Disentanglement Inductive Bias

AutoDAN: Interpretable Gradient-Based Adversarial Attacks on Large Language Models

1 code implementation23 Oct 2023 Sicheng Zhu, Ruiyi Zhang, Bang An, Gang Wu, Joe Barrow, Zichao Wang, Furong Huang, Ani Nenkova, Tong Sun

Safety alignment of Large Language Models (LLMs) can be compromised with manual jailbreak attacks and (automatic) adversarial attacks.

Adversarial Attack Blocking +1

Talking Models: Distill Pre-trained Knowledge to Downstream Models via Interactive Communication

no code implementations4 Oct 2023 Zhe Zhao, Qingyun Liu, Huan Gui, Bang An, Lichan Hong, Ed H. Chi

In this paper, we extend KD with an interactive communication process to help students of downstream tasks learn effectively from pre-trained foundation models.

Decoder Knowledge Distillation +1

AceGPT, Localizing Large Language Models in Arabic

1 code implementation21 Sep 2023 Huang Huang, Fei Yu, Jianqing Zhu, Xuening Sun, Hao Cheng, Dingjie Song, Zhihong Chen, Abdulmohsen Alharthi, Bang An, Juncai He, Ziche Liu, Zhiyi Zhang, Junying Chen, Jianquan Li, Benyou Wang, Lian Zhang, Ruoyu Sun, Xiang Wan, Haizhou Li, Jinchao Xu

This paper is devoted to the development of a localized Large Language Model (LLM) specifically for Arabic, a language imbued with unique cultural characteristics inadequately addressed by current mainstream models.

Instruction Following Language Modeling +3

PerceptionCLIP: Visual Classification by Inferring and Conditioning on Contexts

1 code implementation2 Aug 2023 Bang An, Sicheng Zhu, Michael-Andrei Panaitescu-Liess, Chaithanya Kumar Mummadi, Furong Huang

Inspired by it, we observe that providing CLIP with contextual attributes improves zero-shot image classification and mitigates reliance on spurious features.

Classification Image Classification +4

GFairHint: Improving Individual Fairness for Graph Neural Networks via Fairness Hint

no code implementations25 May 2023 Paiheng Xu, YuHang Zhou, Bang An, Wei Ai, Furong Huang

Given the growing concerns about fairness in machine learning and the impressive performance of Graph Neural Networks (GNNs) on graph data learning, algorithmic fairness in GNNs has attracted significant attention.

Fairness Link Prediction

On the Possibilities of AI-Generated Text Detection

no code implementations10 Apr 2023 Souradip Chakraborty, Amrit Singh Bedi, Sicheng Zhu, Bang An, Dinesh Manocha, Furong Huang

Our work addresses the critical issue of distinguishing text generated by Large Language Models (LLMs) from human-produced text, a task essential for numerous applications.

Text Detection

Transferring Fairness under Distribution Shifts via Fair Consistency Regularization

1 code implementation26 Jun 2022 Bang An, Zora Che, Mucong Ding, Furong Huang

In many real-world applications, however, such an assumption is often violated as previously trained fair models are often deployed in a different environment, and the fairness of such models has been observed to collapse.

Fairness

HintNet: Hierarchical Knowledge Transfer Networks for Traffic Accident Forecasting on Heterogeneous Spatio-Temporal Data

1 code implementation7 Mar 2022 Bang An, Amin Vahedian, Xun Zhou, W. Nick Street, Yanhua Li

However, this problem is challenging due to the spatial heterogeneity of the environment and the sparsity of accidents in space and time.

Management Transfer Learning

Understanding the Generalization Benefit of Model Invariance from a Data Perspective

1 code implementation NeurIPS 2021 Sicheng Zhu, Bang An, Furong Huang

Based on this notion, we refine the generalization bound for invariant models and characterize the suitability of a set of data transformations by the sample covering number induced by transformations, i. e., the smallest size of its induced sample covers.

Generalization Bounds

Adaptive Transfer Learning on Graph Neural Networks

1 code implementation19 Jul 2021 Xueting Han, Zhenhuan Huang, Bang An, Jing Bai

We design an adaptive auxiliary loss weighting model to learn the weights of auxiliary tasks by quantifying the consistency between auxiliary tasks and the target task.

Meta-Learning Multi-Task Learning

Guess First to Enable Better Compression and Adversarial Robustness

no code implementations10 Jan 2020 Sicheng Zhu, Bang An, Shiyu Niu

Machine learning models are generally vulnerable to adversarial examples, which is in contrast to the robustness of humans.

Adversarial Robustness

Cannot find the paper you are looking for? You can Submit a new open access paper.