2 code implementations • 30 Mar 2024 • Zhishang Luo, Truong Son Hy, Puoya Tabaghi, Donghyeon Koh, Michael Defferrard, Elahe Rezaei, Ryan Carey, Rhett Davis, Rajeev Jain, Yusu Wang
Using the input and output data of the tools from past designs, one can attempt to build a machine learning model that predicts the outcome of a design in significantly shorter time than running the tool.
1 code implementation • 2 Mar 2024 • Chenhui Deng, Zichao Yue, Cunxi Yu, Gokce Sarar, Ryan Carey, Rajeev Jain, Zhiru Zhang
In this work we propose HOGA, a novel attention-based model for learning circuit representations in a scalable and generalizable manner.
no code implementations • 31 May 2023 • Ryan Carey, Tom Everitt
How can humans stay in control of advanced artificial intelligence systems?
no code implementations • 5 Jan 2023 • Lewis Hammond, James Fox, Tom Everitt, Ryan Carey, Alessandro Abate, Michael Wooldridge
Regarding question iii), we describe correspondences between causal games and other formalisms, and explain how causal games can be used to answer queries that other causal or game-theoretic models do not support.
no code implementations • 21 Apr 2022 • Sebastian Farquhar, Ryan Carey, Tom Everitt
We then train agents to maximize the causal effect of actions on the expected return which is not mediated by the delicate parts of state, using Causal Influence Diagram analysis.
1 code implementation • 5 Apr 2022 • Animesh Basak Chowdhury, Benjamin Tan, Ryan Carey, Tushit Jain, Ramesh Karri, Siddharth Garg
Generating sub-optimal synthesis transformation sequences ("synthesis recipe") is an important problem in logic synthesis.
no code implementations • 23 Feb 2022 • Chris van Merwijk, Ryan Carey, Tom Everitt
Influence diagrams have recently been used to analyse the safety and fairness properties of AI systems.
no code implementations • 22 Feb 2022 • Carolyn Ashurst, Ryan Carey, Silvia Chiappa, Tom Everitt
In addition to reproducing discriminatory relationships in the training data, machine learning systems can also introduce or amplify discriminatory effects.
no code implementations • 2 Feb 2021 • Tom Everitt, Ryan Carey, Eric Langlois, Pedro A Ortega, Shane Legg
We propose a new graphical criterion for value of control, establishing its soundness and completeness.
no code implementations • 20 Jan 2020 • Ryan Carey, Eric Langlois, Tom Everitt, Shane Legg
Which variables does an agent have an incentive to control with its decision, and which variables does it have an incentive to respond to?
no code implementations • 11 Nov 2019 • Vojtěch Kovařík, Ryan Carey
For some problems, humans may not be able to accurately judge the goodness of AI-proposed solutions.
no code implementations • 19 Sep 2017 • Ryan Carey
We demonstrate this by presenting some Supervised POMDP scenarios in which errors in the parameterized reward function remove the incentive to follow shutdown commands.