no code implementations • 6 Nov 2023 • Ehsan Pajouheshgar, Yitao Xu, Alexander Mordvintsev, Eyvind Niklasson, Tong Zhang, Sabine Süsstrunk
We propose Mesh Neural Cellular Automata (MeshNCA), a method for directly synthesizing dynamic textures on 3D meshes without requiring any UV maps.
no code implementations • 18 Jul 2023 • Ettore Randazzo, Alexander Mordvintsev
Finally, we show how to perform interactive evolution, where the user decides how to evolve a plant model interactively and then deploys it in a larger environment.
2 code implementations • 19 Feb 2023 • Ettore Randazzo, Alexander Mordvintsev, Craig Fouts
Neural Cellular Automata (NCA) models have shown remarkable capacity for pattern formation and complex global behaviors stemming from local coordination.
no code implementations • 6 Feb 2023 • Alexander Mordvintsev, Ettore Randazzo, Eyvind Niklasson
We present a differentiable formulation of abstract chemical reaction networks (CRNs) that can be trained to solve a variety of computational tasks.
1 code implementation • 15 Dec 2022 • Johannes von Oswald, Eyvind Niklasson, Ettore Randazzo, João Sacramento, Alexander Mordvintsev, Andrey Zhmoginov, Max Vladymyrov
We start by providing a simple weight construction that shows the equivalence of data transformations induced by 1) a single linear self-attention layer and by 2) gradient-descent (GD) on a regression loss.
no code implementations • 3 May 2022 • Alexander Mordvintsev, Ettore Randazzo, Craig Fouts
Modeling the ability of multicellular organisms to build and maintain their bodies through local interactions between individual cells (morphogenesis) is a long-standing challenge of developmental biology.
3 code implementations • 26 Nov 2021 • Alexander Mordvintsev, Eyvind Niklasson
We study the problem of example-based procedural texture synthesis using highly compact models.
no code implementations • 22 Jun 2021 • Alexander Mordvintsev, Ettore Randazzo, Eyvind Niklasson
Reaction-Diffusion (RD) systems provide a computational framework that governs many pattern formation processes in nature.
3 code implementations • 15 May 2021 • Alexander Mordvintsev, Eyvind Niklasson, Ettore Randazzo
Neural Cellular Automata (NCA) have shown a remarkable ability to learn the required rules to "grow" images, classify morphologies, segment images, as well as to do general computation such as path-finding.
no code implementations • 11 Aug 2020 • Mark Sandler, Andrey Zhmoginov, Liangcheng Luo, Alexander Mordvintsev, Ettore Randazzo, Blaise Agúera y Arcas
The update rule is applied repeatedly in parallel to a large random subset of cells and after convergence is used to produce segmentation masks that are then back-propagated to learn the optimal update rules using standard gradient descent methods.
2 code implementations • 2 Jul 2020 • Ettore Randazzo, Eyvind Niklasson, Alexander Mordvintsev
We present a novel method for learning the weights of an artificial neural network - a Message Passing Learning Protocol (MPLP).
1 code implementation • Distill 2020 • Chris Olah, Alexander Mordvintsev, Ludwig Schubert
There is a growing sense that neural networks need to be interpretable to humans.
2 code implementations • Distill 2018 • Alexander Mordvintsev, Nicola Pezzotti, Ludwig Schubert, Chris Olah
Typically, we parameterize the input image as the RGB values of each pixel, but that isn’t the only way.
1 code implementation • 28 May 2018 • Nicola Pezzotti, Julian Thijssen, Alexander Mordvintsev, Thomas Hollt, Baldur van Lew, Boudewijn P. F. Lelieveldt, Elmar Eisemann, Anna Vilanova
The t-distributed Stochastic Neighbor Embedding (tSNE) algorithm has become in recent years one of the most used and insightful techniques for the exploratory data analysis of high-dimensional data.
1 code implementation • Distill 2018 • Chris Olah, Arvind Satyanarayan, Ian Johnson, Shan Carter, Ludwig Schubert, Katherine Ye, Alexander Mordvintsev
In this article, we treat existing interpretability methods as fundamental and composable building blocks for rich user interfaces.
2 code implementations • ICCV 2017 • Philip Haeusser, Thomas Frerix, Alexander Mordvintsev, Daniel Cremers
Our training scheme follows the paradigm that in order to effectively derive class labels for the target domain, a network should produce statistically domain invariant embeddings, while minimizing the classification error on the labeled source domain.
Ranked #6 on Domain Adaptation on SYNSIG-to-GTSRB
no code implementations • CVPR 2017 • Philip Haeusser, Alexander Mordvintsev, Daniel Cremers
We demonstrate the capabilities of learning by association on several data sets and show that it can improve performance on classification tasks tremendously by making use of additionally available unlabeled data.
1 code implementation • 3 Jun 2017 • Philip Häusser, Alexander Mordvintsev, Daniel Cremers
We demonstrate the capabilities of learning by association on several data sets and show that it can improve performance on classification tasks tremendously by making use of additionally available unlabeled data.