no code implementations • 11 Feb 2025 • Ido Levy, Orr Paradise, Boaz Carmeli, Ron Meir, Shafi Goldwasser, Yonatan Belinkov
Emergent Communication (EC) provides a unique window into the language systems that emerge autonomously when agents are trained to jointly achieve shared goals.
no code implementations • 13 Jun 2024 • Lior Friedman, Ron Meir
In continual learning, knowledge must be preserved and re-used between tasks, maintaining good transfer to future tasks and minimizing forgetting of previously learned ones.
1 code implementation • 5 May 2024 • Ron Teichner, Ron Meir, Michael Margaliot
Given a time-series of noisy measured outputs of a dynamical system z[k], k=1... N, the Identifying Regulation with Adversarial Surrogates (IRAS) algorithm aims to find a non-trivial first integral of the system, namely, a scalar function g() such that g(z[i]) = g(z[j]), for all i, j.
no code implementations • 17 Mar 2024 • Boaz Carmeli, Yonatan Belinkov, Ron Meir
Artificial agents that learn to communicate in order to accomplish a given task acquire communication protocols that are typically opaque to a human.
no code implementations • 20 Feb 2024 • Omer Cohen, Ron Meir, Nir Weinberger
In the single source case, we propose an elimination learning method, whose risk matches that of a strong-oracle learner.
no code implementations • 3 Feb 2024 • Dror Freirich, Nir Weinberger, Ron Meir
We provide a structural characterization of the DP tradeoff, where the DP function is piecewise linear in the perception index.
no code implementations • 29 Nov 2023 • Ron Teichner, Naama Brenner, Ron Meir
Homeostasis, the ability to maintain a stable internal environment in the face of perturbations, is essential for the functioning of living systems.
no code implementations • 4 Nov 2022 • Boaz Carmeli, Ron Meir, Yonatan Belinkov
The field of emergent communication aims to understand the characteristics of communication as it emerges from artificial agents solving tasks that require information exchange.
1 code implementation • 1 Jul 2022 • Ron Amit, Baruch Epstein, Shay Moran, Ron Meir
We present a PAC-Bayes-style generalization bound which enables the replacement of the KL-divergence with a variety of Integral Probability Metrics (IPM).
no code implementations • 31 May 2022 • Ilya Osadchiy, Kfir Y. Levy, Ron Meir
This solution comprises an inner learner that plays each episode separately, and an outer learner that updates the hyper-parameters of the inner algorithm between the episodes.
1 code implementation • 16 Feb 2022 • Ron Teichner, Ron Meir, Danny Eitan
Our proposed approach consists of three stages: (i) Use the abundant offline data in order to create both a non-causal and a causal estimator for the relevant unmeasured physiological variables.
no code implementations • 12 Jul 2021 • Amit Peleg, Naama Pearl, Ron Meir
In this work we prove, in the context of stochastic linear bandits and Gaussian priors, that as long as the prior is sufficiently close to the true prior, the performance of the applied algorithm is close to that of the algorithm that uses the true prior.
no code implementations • NeurIPS 2021 • Dror Freirich, Tomer Michaeli, Ron Meir
In this paper, we derive a closed form expression for this distortion-perception (DP) function for the mean squared-error (MSE) distortion and the Wasserstein-2 perception index.
no code implementations • 28 Feb 2021 • Oren Peer, Chen Tessler, Nadav Merlis, Ron Meir
Finally, We demonstrate the superior performance of a deep RL variant of EBQL over other deep QL algorithms for a suite of ATARI games.
1 code implementation • ICML 2020 • Ron Amit, Ron Meir, Kamil Ciosek
Specifying a Reinforcement Learning (RL) task involves choosing a suitable planning horizon, which is typically modeled by a discount factor.
1 code implementation • ICML 2020 • Amitay Bar, Ronen Talmon, Ron Meir
Options have been shown to be an effective tool in reinforcement learning, facilitating improved exploration and learning.
no code implementations • 23 May 2019 • Or Raveh, Ron Meir
We develop model free PAC performance guarantees for multiple concurrent MDPs, extending recent works where a single learner interacts with multiple non-interacting agents in a noise free environment.
Multi-agent Reinforcement Learning
reinforcement-learning
+2
no code implementations • 4 Feb 2019 • Baruch Epstein, Ron Meir
Autoencoders are widely used for unsupervised learning and as a regularization scheme in semi-supervised learning.
no code implementations • 6 Aug 2018 • Dror Freirich, Ron Meir, Aviv Tamar
In this formulation, DiRL can be seen as learning a deep generative model of the value distribution, driven by the discrepancy between the distribution of the current value, and the distribution of the sum of current reward and next value.
no code implementations • ICLR 2018 • Ron Amit, Ron Meir
We develop a framework for lifelong learning in deep neural networks that is based on generalization bounds, developed within the PAC-Bayes framework.
no code implementations • ICLR 2018 • Baruch Epstein, Ron Meir, Tomer Michaeli
Ideally one would like to allow both the data for the current task and for previous related tasks to self-organize the learning system in such a way that commonalities and differences between the tasks are learned in a data-driven fashion.
1 code implementation • ICML 2018 • Ron Amit, Ron Meir
In meta-learning an agent extracts knowledge from observed tasks, aiming to facilitate learning of novel future tasks.
no code implementations • 24 Jan 2017 • Alon Hazan, Yuval Harel, Ron Meir
The Human visual perception of the world is of a large fixed image that is highly detailed and sharp.
no code implementations • 12 Sep 2016 • Yuval Harel, Ron Meir, Manfred Opper
The process of dynamic state estimation (filtering) based on point process observations is in general intractable.
no code implementations • NeurIPS 2015 • Yuval Harel, Ron Meir, Manfred Opper
The process of dynamic state estimation (filtering) based on point process observations is in general intractable.
no code implementations • 6 Nov 2015 • Gal Mishne, Ronen Talmon, Ron Meir, Jackie Schiller, Uri Dubin, Ronald R. Coifman
In the wake of recent advances in experimental methods in neuroscience, the ability to record in-vivo neuronal activity from awake animals has become feasible.
no code implementations • 28 Jul 2015 • Yuval Harel, Ron Meir, Manfred Opper
The process of dynamic state estimation (filtering) based on point process observations is in general intractable.
2 code implementations • NeurIPS 2014 • Daniel Soudry, Itay Hubara, Ron Meir
Using online EP and the central limit theorem we find an analytical approximation to the Bayes update of this posterior, as well as the resulting Bayes estimates of the weights and outputs.
no code implementations • NeurIPS 2014 • Alex K. Susemihl, Ron Meir, Manfred Opper
Within the framework of optimal Control Theory, one is usually given a cost function which is minimized by selecting a control law based on the observations.
no code implementations • 27 Jun 2014 • Alex Susemihl, Ron Meir, Manfred Opper
Within the framework of optimal Control Theory, one is usually given a cost function which is minimized by selecting a control law based on the observations.
no code implementations • 7 Oct 2013 • Daniel Soudry, Ron Meir
Significant success has been reported recently using deep neural networks for classification.
no code implementations • NeurIPS 2011 • Alex K. Susemihl, Ron Meir, Manfred Opper
Bayesian filtering of stochastic stimuli has received a great deal of attention re- cently.
no code implementations • NeurIPS 2008 • Dotan D. Castro, Dmitry Volkinshtein, Ron Meir
Actor-critic algorithms for reinforcement learning are achieving renewed popularity due to their good convergence properties in situations where other approaches often fail (e. g., when function approximation is involved).