1 code implementation • 19 Feb 2024 • Fengqing Jiang, Zhangchen Xu, Luyao Niu, Zhen Xiang, Bhaskar Ramasubramanian, Bo Li, Radha Poovendran
In this paper, we propose a novel ASCII art-based jailbreak attack and introduce a comprehensive benchmark Vision-in-Text Challenge (ViTC) to evaluate the capabilities of LLMs in recognizing prompts that cannot be solely interpreted by semantics.
no code implementations • 12 Feb 2024 • Dinuka Sahabandu, Xiaojun Xu, Arezoo Rajabi, Luyao Niu, Bhaskar Ramasubramanian, Bo Li, Radha Poovendran
We propose and analyze an adaptive adversary that can retrain a Trojaned DNN and is also aware of SOTA output-based Trojaned model detectors.
no code implementations • 2 Feb 2024 • Arezoo Rajabi, Reeya Pimple, Aiswarya Janardhanan, Surudhi Asokraj, Bhaskar Ramasubramanian, Radha Poovendran
However, the suitability of TL as a solution to reduce vulnerability of overfitted DNNs to privacy attacks is unexplored.
1 code implementation • 20 Jan 2024 • Zhen Xiang, Fengqing Jiang, Zidi Xiong, Bhaskar Ramasubramanian, Radha Poovendran, Bo Li
Moreover, we show that LLMs endowed with stronger reasoning capabilities exhibit higher susceptibility to BadChain, exemplified by a high average attack success rate of 97. 0% across the six benchmark tasks on GPT-4.
1 code implementation • 30 Aug 2023 • Arezoo Rajabi, Surudhi Asokraj, Fengqing Jiang, Luyao Niu, Bhaskar Ramasubramanian, Jim Ritcey, Radha Poovendran
An adversary carrying out a backdoor attack embeds a predefined perturbation called a trigger into a small subset of input samples and trains the DNN such that the presence of the trigger in the input results in an adversary-desired output class.
no code implementations • 4 Apr 2023 • Abdullah Al Maruf, Luyao Niu, Bhaskar Ramasubramanian, Andrew Clark, Radha Poovendran
We then propose a distributed MARL algorithm called the CVaR QD-Learning algorithm, and establish that value functions of individual agents reaches consensus.
no code implementations • 3 Dec 2022 • Arezoo Rajabi, Dinuka Sahabandu, Luyao Niu, Bhaskar Ramasubramanian, Radha Poovendran
Overfitted models have been shown to be susceptible to query-based attacks such as membership inference attacks (MIAs).
no code implementations • 13 Jul 2022 • Dinuka Sahabandu, Arezoo Rajabi, Luyao Niu, Bo Li, Bhaskar Ramasubramanian, Radha Poovendran
The results show that (i) with Submodular Trojan algorithm, the adversary needs to embed a Trojan trigger into a very small fraction of samples to achieve high accuracy on both Trojan and clean samples, and (ii) the MM Trojan algorithm yields a trained Trojan model that evades detection with probability 1.
no code implementations • 25 Mar 2022 • Arezoo Rajabi, Bhaskar Ramasubramanian, Radha Poovendran
We term this category of attacks multi-target backdoor attacks.
no code implementations • 18 Mar 2022 • Arezoo Rajabi, Bhaskar Ramasubramanian, Abdullah Al Maruf, Radha Poovendran
Through empirical evaluations, we highlight a privacy-utility tradeoff and demonstrate that the RL agent is able to learn behaviors that are aligned with that of a human user in the same environment in a privacy-preserving manner
1 code implementation • 19 Feb 2022 • Baicen Xiao, Bhaskar Ramasubramanian, Radha Poovendran
We design two algorithms- Shaping Advice in Single-agent reinforcement learning (SAS) and Shaping Advice in Multi-agent reinforcement learning (SAM).
Multi-agent Reinforcement Learning reinforcement-learning +1
1 code implementation • 12 Jan 2022 • Baicen Xiao, Bhaskar Ramasubramanian, Radha Poovendran
In this paper, we introduce Agent-Temporal Attention for Reward Redistribution in Episodic Multi-Agent Reinforcement Learning (AREL) to address these two challenges.
Multi-agent Reinforcement Learning reinforcement-learning +2
no code implementations • 2 Sep 2021 • Bhaskar Ramasubramanian, M. A. Rajan, M. Girish Chandra, Rance Cleaveland, Steven I. Marcus
The resilience of cyberphysical systems to denial-of-service (DoS) and integrity attacks is studied in this paper.
no code implementations • 29 Mar 2021 • Bhaskar Ramasubramanian, Luyao Niu, Andrew Clark, Radha Poovendran
In this paper, we consider a setting where an autonomous agent has to learn behaviors in an unknown environment.
1 code implementation • 29 Mar 2021 • Baicen Xiao, Bhaskar Ramasubramanian, Radha Poovendran
We observe that using SAM results in agents learning policies to complete tasks faster, and obtain higher rewards than: i) using sparse rewards alone; ii) a state-of-the-art reward redistribution method.
Multi-agent Reinforcement Learning reinforcement-learning +1
no code implementations • 20 Sep 2020 • Bhaskar Ramasubramanian, Baicen Xiao, Linda Bushnell, Radha Poovendran
We propose an iterative approach to the synthesis of the controller by solving a modified discrete-time Riccati equation.
no code implementations • 19 Jan 2020 • Baicen Xiao, Qifan Lu, Bhaskar Ramasubramanian, Andrew Clark, Linda Bushnell, Radha Poovendran
The output of the feedback neural network is converted to a shaping reward that is augmented to the reward provided by the environment.
no code implementations • 20 Jul 2019 • Baicen Xiao, Bhaskar Ramasubramanian, Andrew Clark, Hannaneh Hajishirzi, Linda Bushnell, Radha Poovendran
This paper augments the reward received by a reinforcement learning agent with potential functions in order to help the agent learn (possibly stochastic) optimal policies.