1 code implementation • 18 Jan 2022 • Tong Mu, Stephan Zheng, Alexander Trott
In a sequential setting with multiple Agents, RIRL shows opposing consequences of the Principal's inattention to different information channels: 1) inattention to Agents' outputs closes wage gaps based on ability differences; and 2) inattention to Agents' efforts induces a social dilemma dynamic in which Agents work harder, but essentially for free.
Deep Reinforcement Learning
Multi-agent Reinforcement Learning
no code implementations • 3 Jan 2022 • Michael Curry, Alexander Trott, Soham Phade, Yu Bai, Stephan Zheng
We validate the learned solutions are $\epsilon$-meta-equilibria through best-response analyses, show that they align with economic intuitions, and show our approach can learn a spectrum of qualitatively distinct $\epsilon$-meta-equilibria in open RBC models.
Deep Reinforcement Learning
Multi-agent Reinforcement Learning
+2
1 code implementation • 6 Aug 2021 • Alexander Trott, Sunil Srinivasa, Douwe van der Wal, Sebastien Haneuse, Stephan Zheng
Here we show that the AI Economist framework enables effective, flexible, and interpretable policy design using two-level reinforcement learning (RL) and data-driven simulations.
1 code implementation • 5 Aug 2021 • Stephan Zheng, Alexander Trott, Sunil Srinivasa, David C. Parkes, Richard Socher
Here we show that machine-learning-based economic simulation is a powerful policy and mechanism design framework to overcome these limitations.
2 code implementations • 28 Apr 2020 • Stephan Zheng, Alexander Trott, Sunil Srinivasa, Nikhil Naik, Melvin Gruesbeck, David C. Parkes, Richard Socher
In experiments conducted on MTurk, an AI tax policy provides an equality-productivity trade-off that is similar to that provided by the Saez framework along with higher inverse-income weighted social welfare.
1 code implementation • ICML 2020 • Víctor Campos, Alexander Trott, Caiming Xiong, Richard Socher, Xavier Giro-i-Nieto, Jordi Torres
We perform an extensive evaluation of skill discovery methods on controlled environments and show that EDL offers significant advantages, such as overcoming the coverage problem, reducing the dependence of learned skills on the initial state, and allowing the user to define a prior over which behaviors should be learned.
1 code implementation • NeurIPS 2019 • Alexander Trott, Stephan Zheng, Caiming Xiong, Richard Socher
For instance, in tasks where the agent must achieve some goal state, simple distance-to-goal reward shaping often fails, as it renders learning vulnerable to local optima.
no code implementations • ICLR 2019 • Hao Liu, Alexander Trott, Richard Socher, Caiming Xiong
We propose a novel method called competitive experience replay, which efficiently supplements a sparse reward by placing learning in the context of an exploration competition between a pair of agents.
no code implementations • ICLR 2018 • Alexander Trott, Caiming Xiong, Richard Socher
Questions that require counting a variety of objects in images remain a major challenge in visual question answering (VQA).