1 code implementation • 23 Dec 2024 • Akarsh Kumar, Chris Lu, Louis Kirsch, Yujin Tang, Kenneth O. Stanley, Phillip Isola, David Ha
With the recent Nobel Prize awarded for radical advances in protein discovery, foundation models (FMs) for exploring large combinatorial spaces promise to revolutionize many scientific fields.
no code implementations • 7 Nov 2024 • Usman Anwar, Johannes von Oswald, Louis Kirsch, David Krueger, Spencer Frei
This work investigates the vulnerability of in-context learning in transformers to \textit{hijacking attacks} focusing on the setting of linear regression tasks.
1 code implementation • 6 May 2024 • Aditya A. Ramesh, Kenny Young, Louis Kirsch, Jürgen Schmidhuber
Temporal credit assignment in reinforcement learning is challenging due to delayed and stochastic outcomes.
2 code implementations • 26 Feb 2024 • Mingchen Zhuge, Wenyi Wang, Louis Kirsch, Francesco Faccio, Dmitrii Khizbullin, Jürgen Schmidhuber
Various human-designed prompt engineering techniques have been proposed to improve problem solvers based on Large Language Models (LLMs), yielding many disparate code bases.
1 code implementation • 8 Feb 2024 • Matthew Thomas Jackson, Chris Lu, Louis Kirsch, Robert Tjarko Lange, Shimon Whiteson, Jakob Nicolaus Foerster
We propose a simple augmentation to two existing objective discovery approaches that allows the discovered algorithm to dynamically update its objective function throughout the agent's training procedure, resulting in expressive schedules and increased generalization across different training horizons.
1 code implementation • 20 Sep 2023 • Aleksandar Stanić, Dylan Ashley, Oleg Serikov, Louis Kirsch, Francesco Faccio, Jürgen Schmidhuber, Thomas Hofmann, Imanol Schlag
We introduce an experimental protocol that enables model comparisons based on equivalent compute, measured in accelerator hours.
no code implementations • 26 May 2023 • Mingchen Zhuge, Haozhe Liu, Francesco Faccio, Dylan R. Ashley, Róbert Csordás, Anand Gopalakrishnan, Abdullah Hamdi, Hasan Abed Al Kader Hammoud, Vincent Herrmann, Kazuki Irie, Louis Kirsch, Bing Li, Guohao Li, Shuming Liu, Jinjie Mai, Piotr Piękos, Aditya Ramesh, Imanol Schlag, Weimin Shi, Aleksandar Stanić, Wenyi Wang, Yuhui Wang, Mengmeng Xu, Deng-Ping Fan, Bernard Ghanem, Jürgen Schmidhuber
What should be the social structure of an NLSOM?
no code implementations • 18 May 2023 • Samuel Schmidgall, Jascha Achterberg, Thomas Miconi, Louis Kirsch, Rojin Ziaei, S. Pardis Hajiseyedrazi, Jason Eshraghian
Artificial neural networks (ANNs) have emerged as an essential tool in machine learning, achieving remarkable success across diverse domains, including image and speech generation, game playing, and robotics.
no code implementations • 29 Dec 2022 • Louis Kirsch, Jürgen Schmidhuber
We discuss the relationship of such systems to in-context and memory-based meta learning and show that self-referential neural networks require functionality to be reused in the form of parameter sharing.
no code implementations • 29 Dec 2022 • Vincent Herrmann, Louis Kirsch, Jürgen Schmidhuber
There are two important things in science: (A) Finding answers to given questions, and (B) Coming up with good questions.
no code implementations • 8 Dec 2022 • Louis Kirsch, James Harrison, Jascha Sohl-Dickstein, Luke Metz
We further show that the capabilities of meta-trained algorithms are bottlenecked by the accessible state size (memory) determining the next prediction, unlike standard models which are thought to be bottlenecked by parameter count.
1 code implementation • 18 Nov 2022 • Aditya Ramesh, Louis Kirsch, Sjoerd van Steenkiste, Jürgen Schmidhuber
Furthermore, RC-GVF significantly outperforms previous methods in the absence of ground-truth episodic counts in the partially observable MiniGrid environments.
1 code implementation • 4 Nov 2022 • Kenny Young, Aditya Ramesh, Louis Kirsch, Jürgen Schmidhuber
First, we provide a simple theorem motivating how learning a model as an intermediate step can narrow down the set of possible value functions more than learning a value function directly from data using the Bellman equation.
Model-based Reinforcement Learning reinforcement-learning +2
1 code implementation • 4 Jul 2022 • Francesco Faccio, Vincent Herrmann, Aditya Ramesh, Louis Kirsch, Jürgen Schmidhuber
A form of weight-sharing HyperNetworks and policy embeddings scales our method to generate deep NNs.
no code implementations • 22 Sep 2021 • Louis Kirsch, Sebastian Flennerhag, Hado van Hasselt, Abram Friesen, Junhyuk Oh, Yutian Chen
We show that a recent successful meta RL approach that meta-learns an objective for backpropagation-based learning exhibits certain symmetries (specifically the reuse of the learning rule, and invariance to input and output permutations) that are not present in typical black-box meta RL systems.
no code implementations • NeurIPS 2021 • Louis Kirsch, Jürgen Schmidhuber
Many concepts have been proposed for meta learning with neural networks (NNs), e. g., NNs that learn to reprogram fast weights, Hebbian plasticity, learned learning rules, and meta recurrent NNs.
1 code implementation • ICLR 2021 • Francesco Faccio, Louis Kirsch, Jürgen Schmidhuber
We introduce a class of value functions called Parameter-Based Value Functions (PBVFs) whose inputs include the policy parameters.
no code implementations • ICLR 2020 • Louis Kirsch, Sjoerd van Steenkiste, Jürgen Schmidhuber
Biological evolution has distilled the experiences of many learners into the general learning algorithms of humans.
no code implementations • 12 Feb 2019 • Julius Kunze, Louis Kirsch, Hippolyt Ritter, David Barber
Variational inference with a factorized Gaussian posterior estimate is a widely used approach for learning parameters and hidden variables.
no code implementations • NeurIPS 2018 • Louis Kirsch, Julius Kunze, David Barber
Scaling model capacity has been vital in the success of deep learning.
no code implementations • 27 Sep 2018 • Julius Kunze, Louis Kirsch, Hippolyt Ritter, David Barber
We propose Noisy Information Bottlenecks (NIB) to limit mutual information between learned parameters and the data through noise.
3 code implementations • WS 2017 • Julius Kunze, Louis Kirsch, Ilia Kurenkov, Andreas Krug, Jens Johannsmeier, Sebastian Stober
End-to-end training of automated speech recognition (ASR) systems requires massive data and compute resources.