#### Filter Results:

- Full text PDF available (12)

#### Publication Year

2008

2017

- This year (3)
- Last 5 years (9)
- Last 10 years (13)

#### Publication Type

#### Co-author

#### Journals and Conferences

#### Brain Region

#### Cell Type

#### Data Set Used

#### Key Phrases

#### Method

#### Organism

Learn More

- Bernhard Nessler, Michael Pfeiffer, Lars Buesing, Wolfgang Maass
- PLoS Computational Biology
- 2013

The principles by which networks of neurons compute, and how spike-timing dependent plasticity (STDP) of synaptic weights generates and maintains their computational function, are unknown. Preceding work has shown that soft winner-take-all (WTA) circuits, where pyramidal neurons inhibit each other via interneurons, are a common motif of cortical… (More)

- Lars Buesing, Johannes Bill, Bernhard Nessler, Wolfgang Maass
- PLoS Computational Biology
- 2011

The organization of computations in networks of spiking neurons in the brain is still largely unknown, in particular in view of the inherently stochastic features of their firing activity and the experimentally observed trial-to-trial variability of neural systems in the brain. In principle there exists a powerful computational framework for stochastic… (More)

- Bernhard Nessler, Michael Pfeiffer, Wolfgang Maass
- NIPS
- 2009

The principles by which spiking neurons contribute to the astounding computational power of generic cortical microcircuits, and how spike-timing-dependent plasticity (STDP) of synaptic weights could generate and maintain this computational function, are unknown. We show here that STDP, in conjunction with a stochastic soft winner-take-all (WTA) circuit,… (More)

- David Kappel, Bernhard Nessler, Wolfgang Maass
- PLoS Computational Biology
- 2014

In order to cross a street without being run over, we need to be able to extract very fast hidden causes of dynamically changing multi-modal sensory stimuli, and to predict their future evolution. We show here that a generic cortical microcircuit motif, pyramidal cells with lateral excitation and inhibition, provides the basis for this difficult but… (More)

- Stefan Habenschuss, Johannes Bill, Bernhard Nessler
- NIPS
- 2012

Recent spiking network models of Bayesian inference and unsupervised learning frequently assume either inputs to arrive in a special format or employ complex computations in neuronal activation functions and synaptic plasticity rules. Here we show in a rigorous mathematical treatment how homeostatic processes, which have previously received little attention… (More)

- Michael Treml, Jose A. Arjona-Medina, +9 authors Sepp Hochreiter
- 2016

Deep learning has considerably improved semantic image segmentation. However, its high accuracy is traded against larger computational costs which makes it unsuitable for embedded devices in self-driving cars. We propose a novel deep network architecture for image segmentation that keeps the high accuracy while being efficient enough for embedded devices.… (More)

- Michael Pfeiffer, Bernhard Nessler, Rodney J. Douglas, Wolfgang Maass
- Neural Computation
- 2010

We introduce a framework for decision making in which the learning of decision making is reduced to its simplest and biologically most plausible form: Hebbian learning on a linear neuron. We cast our Bayesian-Hebb learning rule as reinforcement learning in which certain decisions are rewarded and prove that each synaptic weight will on average converge… (More)

- Bernhard Nessler, Michael Pfeiffer, Wolfgang Maass
- NIPS
- 2008

Uncertainty is omnipresent when we perceive or interact with our environment, and the Bayesian framework provides computational methods for dealing with it. Mathematical models for Bayesian decision making typically require datastructures that are hard to implement in neural networks. This article shows that even the simplest and experimentally best… (More)

- Christoph Hartmann, Andreea Lazar, Bernhard Nessler, Jochen Triesch
- PLoS Computational Biology
- 2015

Even in the absence of sensory stimulation the brain is spontaneously active. This background "noise" seems to be the dominant cause of the notoriously high trial-to-trial variability of neural recordings. Recent experimental observations have extended our knowledge of trial-to-trial variability and spontaneous activity in several directions: 1.… (More)

Generative Adversarial Networks (GANs) excel at creating realistic images with complex models for which maximum likelihood is infeasible. However, the convergence of GAN training has still not been proved. We propose a two time-scale update rule (TTUR) for training GANs with stochastic gradient descent that has an individual learning rate for both the… (More)