Search Results for author: Goran Radanovic

Found 23 papers, 7 papers with code

Reward Design for Justifiable Sequential Decision-Making

1 code implementation24 Feb 2024 Aleksa Sukovic, Goran Radanovic

In this work, we propose the use of a debate-based reward model for reinforcement learning agents, where the outcome of a zero-sum debate game quantifies the justifiability of a decision in a particular state.

Decision Making

Agent-Specific Effects: A Causal Effect Propagation Analysis in Multi-Agent MDPs

no code implementations17 Oct 2023 Stelios Triantafyllou, Aleksa Sukovic, Debmalya Mandal, Goran Radanovic

These challenges are particularly prominent in the context of multi-agent sequential decision-making, where the causal effect of an agent's action on the outcome depends on how other agents respond to that action.

counterfactual Decision Making +1

Markov Decision Processes with Time-Varying Geometric Discounting

no code implementations19 Jul 2023 Jiarui Gan, Annika Hennes, Rupak Majumdar, Debmalya Mandal, Goran Radanovic

We take a game-theoretic perspective -- whereby each time step is treated as an independent decision maker with their own (fixed) discount factor -- and we study the subgame perfect equilibrium (SPE) of the resulting game as well as the related algorithmic problems.

Sequential Principal-Agent Problems with Communication: Efficient Computation and Learning

no code implementations6 Jun 2023 Jiarui Gan, Rupak Majumdar, Debmalya Mandal, Goran Radanovic

In this model, the principal and the agent interact in a stochastic environment, and each is privy to observations about the state not available to the other.

Decision Making

Learning Embeddings for Sequential Tasks Using Population of Agents

1 code implementation5 Jun 2023 Mridul Mahajan, Georgios Tzannetos, Goran Radanovic, Adish Singla

We present an information-theoretic framework to learn fixed-dimensional embeddings for tasks in reinforcement learning.

Decision Making

Implicit Poisoning Attacks in Two-Agent Reinforcement Learning: Adversarial Policies for Training-Time Attacks

1 code implementation27 Feb 2023 Mohammad Mohammadi, Jonathan Nöther, Debmalya Mandal, Adish Singla, Goran Radanovic

In this paper, we study targeted poisoning attacks in a two-agent setting where an attacker implicitly poisons the effective environment of one of the agents by modifying the policy of its peer.

Online Reinforcement Learning with Uncertain Episode Lengths

no code implementations7 Feb 2023 Debmalya Mandal, Goran Radanovic, Jiarui Gan, Adish Singla, Rupak Majumdar

We show that minimizing regret with this new general discounting is equivalent to minimizing regret with uncertain episode lengths.

reinforcement-learning Reinforcement Learning (RL)

Performative Reinforcement Learning

no code implementations30 Jun 2022 Debmalya Mandal, Stelios Triantafyllou, Goran Radanovic

We introduce the framework of performative reinforcement learning where the policy chosen by the learner affects the underlying reward and transition dynamics of the environment.

reinforcement-learning Reinforcement Learning (RL)

Actual Causality and Responsibility Attribution in Decentralized Partially Observable Markov Decision Processes

no code implementations1 Apr 2022 Stelios Triantafyllou, Adish Singla, Goran Radanovic

Responsibility attribution is complementary and aims to identify the extent to which decision makers (agents) are responsible for this outcome.

Decision Making Decision Making Under Uncertainty

Admissible Policy Teaching through Reward Design

no code implementations6 Jan 2022 Kiarash Banihashem, Adish Singla, Jiarui Gan, Goran Radanovic

This problem can be viewed as a dual to the problem of optimal reward poisoning attacks: instead of forcing an agent to adopt a specific policy, the reward designer incentivizes an agent to avoid taking actions that are inadmissible in certain states.

Explicable Reward Design for Reinforcement Learning Agents

1 code implementation NeurIPS 2021 Rati Devidze, Goran Radanovic, Parameswaran Kamalaruban, Adish Singla

By being explicable, we seek to capture two properties: (a) informativeness so that the rewards speed up the agent's convergence, and (b) sparseness as a proxy for ease of interpretability of the rewards.

Informativeness reinforcement-learning +1

On Blame Attribution for Accountable Multi-Agent Sequential Decision Making

no code implementations NeurIPS 2021 Stelios Triantafyllou, Adish Singla, Goran Radanovic

We formalize desirable properties of blame attribution in the setting of interest, and we analyze the relationship between these properties and the studied blame attribution methods.

Decision Making Fairness

Reinforcement Learning for Education: Opportunities and Challenges

no code implementations15 Jul 2021 Adish Singla, Anna N. Rafferty, Goran Radanovic, Neil T. Heffernan

This survey article has grown out of the RL4ED workshop organized by the authors at the Educational Data Mining (EDM) 2021 conference.

reinforcement-learning Reinforcement Learning (RL)

Defense Against Reward Poisoning Attacks in Reinforcement Learning

no code implementations10 Feb 2021 Kiarash Banihashem, Adish Singla, Goran Radanovic

As a threat model, we consider attacks that minimally alter rewards to make the attacker's target policy uniquely optimal under the poisoned rewards, with the optimality gap specified by an attack parameter.

reinforcement-learning Reinforcement Learning (RL)

Policy Teaching in Reinforcement Learning via Environment Poisoning Attacks

no code implementations21 Nov 2020 Amin Rakhsha, Goran Radanovic, Rati Devidze, Xiaojin Zhu, Adish Singla

We provide lower/upper bounds on the attack cost, and instantiate our attacks in two settings: (i) an offline setting where the agent is doing planning in the poisoned environment, and (ii) an online setting where the agent is learning a policy with poisoned feedback.

reinforcement-learning Reinforcement Learning (RL)

Policy Teaching via Environment Poisoning: Training-time Adversarial Attacks against Reinforcement Learning

1 code implementation ICML 2020 Amin Rakhsha, Goran Radanovic, Rati Devidze, Xiaojin Zhu, Adish Singla

We study a security threat to reinforcement learning where an attacker poisons the learning environment to force the agent into executing a target policy chosen by the attacker.

reinforcement-learning Reinforcement Learning (RL)

Learning to Collaborate in Markov Decision Processes

no code implementations23 Jan 2019 Goran Radanovic, Rati Devidze, David C. Parkes, Adish Singla

We consider a two-agent MDP framework where agents repeatedly solve a task in a collaborative setting.

Multi-View Decision Processes: The Helper-AI Problem

no code implementations NeurIPS 2017 Christos Dimitrakakis, David C. Parkes, Goran Radanovic, Paul Tylkin

We consider a two-player sequential game in which agents have the same reward function but may disagree on the transition probabilities of an underlying Markovian model of the world.

Calibrated Fairness in Bandits

no code implementations6 Jul 2017 Yang Liu, Goran Radanovic, Christos Dimitrakakis, Debmalya Mandal, David C. Parkes

In addition, we define the {\em fairness regret}, which corresponds to the degree to which an algorithm is not calibrated, where perfect calibration requires that the probability of selecting an arm is equal to the probability with which the arm has the best quality realization.

Decision Making Fairness +1

Bayesian fairness

no code implementations31 May 2017 Christos Dimitrakakis, Yang Liu, David Parkes, Goran Radanovic

We consider the problem of how decision making can be fair when the underlying probabilistic model of the world is not known with certainty.

BIG-bench Machine Learning Decision Making +1

Cannot find the paper you are looking for? You can Submit a new open access paper.