Search Results for author: Christian K. Machens

Found 11 papers, 3 papers with code

Approximating nonlinear functions with latent boundaries in low-rank excitatory-inhibitory spiking networks

1 code implementation18 Jul 2023 William F. Podlaski, Christian K. Machens

Overall, our work proposes a new perspective on spiking networks that may serve as a starting point for a mechanistic understanding of biological spike-based computation.

Compact task representations as a normative model for higher-order brain activity

no code implementations NeurIPS 2020 Severin Berger, Christian K. Machens

More specifically, we focus on MDPs whose state is based on action-observation histories, and we show how to compress the state space such that unnecessary redundancy is eliminated, while task-relevant information is preserved.

Understanding spiking networks through convex optimization

1 code implementation NeurIPS 2020 Allan Mancoo, Sander Keemink, Christian K. Machens

Here we turn these findings around and show that virtually all inhibition-dominated SNNs can be understood through the lens of convex optimization, with network connectivity, timescales, and firing thresholds being intricately linked to the parameters of underlying convex optimization problems.

Open-Ended Question Answering

Extracting Latent Structure From Multiple Interacting Neural Populations

no code implementations NeurIPS 2014 Joao Semedo, Amin Zandvakili, Adam Kohn, Christian K. Machens, Byron M. Yu

Developments in neural recording technology are rapidly enabling the recording of populations of neurons in multiple brain areas simultaneously, as well as the identification of the types of neurons being recorded (e. g., excitatory vs. inhibitory).

Unsupervised learning of an efficient short-term memory network

no code implementations NeurIPS 2014 Pietro Vertechi, Wieland Brendel, Christian K. Machens

Specifically, we show how these networks can learn to efficiently represent their present and past inputs, based on local learning rules only.

Firing rate predictions in optimal balanced networks

no code implementations NeurIPS 2013 David G. Barrett, Sophie Denève, Christian K. Machens

This is an important problem because firing rates are one of the most important measures of network activity, in both the study of neural computation and neural network dynamics.

Learning optimal spike-based representations

no code implementations NeurIPS 2012 Ralph Bourdoukan, David Barrett, Sophie Deneve, Christian K. Machens

We define a loss function for the quality of the population read-out and derive the dynamical equations for both neurons and synapses from the requirement to minimize this loss.

Demixed Principal Component Analysis

no code implementations NeurIPS 2011 Wieland Brendel, Ranulfo Romo, Christian K. Machens

Standard dimensionality reduction techniques such as principal component analysis (PCA) can provide a succinct and complete description of the data, but the description is constructed independent of the relevant task variables and is often hard to interpret.

Dimensionality Reduction

Linear readout from a neural population with partial correlation data

no code implementations NeurIPS 2010 Adrien Wohrer, Ranulfo Romo, Christian K. Machens

We compare the discrimination thresholds read out from the population of recorded neurons with the discrimination threshold of the monkey and show that our method predicts different results than simpler, average schemes of noise correlations."

Cannot find the paper you are looking for? You can Submit a new open access paper.