1 code implementation • 18 Dec 2023 • Yotam Amitai, Yael Septon, Ofra Amir
Explainable reinforcement learning (XRL) methods aim to help elucidate agent policies and decision-making processes.
no code implementations • 24 Jan 2023 • Yotam Amitai, Guy Avni, Ofra Amir
As reinforcement learning methods increasingly amass accomplishments, the need for comprehending their solutions becomes more crucial.
no code implementations • 21 Oct 2022 • Yael Septon, Tobias Huber, Elisabeth André, Ofra Amir
Methods that help users understand the behavior of such agents can roughly be divided into local explanations that analyze specific decisions of the agents and global explanations that convey the general strategy of the agents.
1 code implementation • 14 Mar 2022 • Hana Matatov, Mor Naaman, Ofra Amir
The massive spread of visual content through the web and social media poses both challenges and opportunities.
no code implementations • 5 Feb 2021 • Yotam Amitai, Ofra Amir
In this paper, we propose a novel method for generating dependent and contrastive summaries that emphasize the differences between agent policies by identifying states in which the agents disagree on the best course of action.
1 code implementation • 20 Jan 2021 • Anton Abilov, Yiqing Hua, Hana Matatov, Ofra Amir, Mor Naaman
The wide spread of unfounded election fraud claims surrounding the U. S. 2020 election had resulted in undermining of trust in the election, culminating in violence inside the U. S. capitol.
Social and Information Networks
1 code implementation • 2 Dec 2020 • Roee Shraga, Ofra Amir, Avigdor Gal
Matching is a task at the heart of any data integration process, aimed at identifying correspondences among data elements.
1 code implementation • 18 May 2020 • Tobias Huber, Katharina Weitz, Elisabeth André, Ofra Amir
Specifically, we augment strategy summaries that extract important trajectories of states from simulations of the agent with saliency maps which show what information the agent attends to.
1 code implementation • 30 May 2019 • Isaac Lage, Daphna Lifschitz, Finale Doshi-Velez, Ofra Amir
We introduce an imitation learning-based approach to policy summarization; we demonstrate through computational simulations that a mismatch between the model used to extract a summary and the model used to reconstruct the policy results in worse reconstruction quality; and we demonstrate through a human-subject study that people use different models to reconstruct policies in different contexts, and that matching the summary extraction model to these can improve performance.
no code implementations • 2 May 2019 • Reshef Meir, Ofra Amir, Omer Ben-Porat, Tsviel Ben-Shabat, Gal Cohensius, Lirong Xia
Truth discovery is a general name for a broad range of statistical methods aimed to extract the correct answers to questions, based on multiple answers coming from noisy sources.