no code implementations • 29 Jun 2024 • Robert Loftin, Saptarashmi Bandyopadhyay, Mustafa Mert Çelikok
Artificially intelligent agents deployed in the real-world will require the ability to reliably \textit{cooperate} with humans (as well as other, heterogeneous AI agents).
no code implementations • 26 Jul 2023 • Bram Renting, Phillip Wozny, Robert Loftin, Claudia Wieners, Erman Acar
We present a critical analysis of the simulation framework RICE-N, an integrated assessment model (IAM) for evaluating the impacts of climate change on the economy.
no code implementations • 26 Jul 2023 • Phillip Wozny, Bram Renting, Robert Loftin, Claudia Wieners, Erman Acar
As our submission for track three of the AI for Global Climate Cooperation (AI4GCC) competition, we propose a negotiation protocol for use in the RICE-N climate-economic simulation.
no code implementations • 29 May 2023 • Robert Loftin, Mustafa Mert Çelikok, Frans A. Oliehoek
Multiagent systems deployed in the real world need to cooperate with other agents (including humans) nearly as effectively as these agents cooperate with one another.
1 code implementation • 7 Feb 2023 • Robert Loftin, Mustafa Mert Çelikok, Herke van Hoof, Samuel Kaski, Frans A. Oliehoek
A natural solution concept in such settings is the Stackelberg equilibrium, in which the ``leader'' agent selects the strategy that maximizes its own payoff given that the ``follower'' agent will choose their best response to this strategy.
no code implementations • 20 Jun 2022 • Robert Loftin, Frans A. Oliehoek
Learning to cooperate with other agents is challenging when those agents also possess the ability to adapt to our own behavior.
1 code implementation • 30 Jul 2021 • Robert Loftin, Aadirupa Saha, Sam Devlin, Katja Hofmann
High sample complexity remains a barrier to the application of reinforcement learning (RL), particularly in multi-agent systems.
1 code implementation • NeurIPS 2019 • Kamil Ciosek, Quan Vuong, Robert Loftin, Katja Hofmann
To address both of these phenomena, we introduce a new algorithm, Optimistic Actor Critic, which approximates a lower and upper confidence bound on the state-action value function.
no code implementations • 28 Oct 2019 • Kamil Ciosek, Quan Vuong, Robert Loftin, Katja Hofmann
To address both of these phenomena, we introduce a new algorithm, Optimistic Actor Critic, which approximates a lower and upper confidence bound on the state-action value function.
no code implementations • 19 Jul 2019 • Robert Loftin, Bei Peng, Matthew E. Taylor, Michael L. Littman, David L. Roberts
In order for robots and other artificial agents to efficiently learn to perform useful tasks defined by an end user, they must understand not only the goals of those tasks, but also the structure and dynamics of that user's environment.
no code implementations • ICML 2017 • James MacGlashan, Mark K. Ho, Robert Loftin, Bei Peng, Guan Wang, David Roberts, Matthew E. Taylor, Michael L. Littman
This paper investigates the problem of interactively learning behaviors communicated by a human teacher using positive and negative feedback.