Profile picture placeholder

Wulfram Gerstner

Expertise


Computational Neuroscience, Theoretical Neuroscience

Current Work

- Neuronal Dynamics in large recurrent networks, in particular networks of spiking neurons
- biologically plausible learning rules

 Wulfram Gerstner is Director of the Laboratory of Computational Neuroscience LCN at the EPFL. His research in computational neuroscience concentrates on models of spiking neurons and spike-timing dependent plasticity, on the problem of neuronal coding and neural dynamics in single neurons and populations, as well as on the link between biologically plausible learning rules and behavioral manifestations of learning. He teaches courses for Physicists, Computer Scientists, Mathematicians, and Life Scientists at the EPFL.


Curriculum vitae

 After studies of Physics in Tübingen and at the Ludwig-Maximilians-University Munich (Master 1989), Wulfram Gerstner spent a year as a visiting researcher in Berkeley. He received his PhD in theoretical physics from the Technical University Munich in 1993 with a thesis on associative memory and dynamics in networks of spiking neurons. After short postdoctoral stays at Brandeis University and the Technical University of Munich, he joined the EPFL in 1996 as assistant professor. Promoted to Associate Professor with tenure in February 2001, he is since August 2006 a full professor with double appointment in the School of Computer and Communication Sciences and the School of Life Sciences. Wulfram Gerstner has been invited speaker at numerous international conferences and workshops. He has served on the editorial board of the Journal of Neuroscience, `Network: Computation in Neural Systems', `Journal of Computational Neuroscience', and `Science'.  

Awards

Valentino Braitenberg Award for Computational Neuroscience 2018

Bernstein Network Computational Neuroscience

2018

Member of the Academy of Ccience and Literature, Mainz, Germany

Academy of Science and Literature, Mainz, Germany

2019

Selected publications

Novelty as a drive of human exploration in complex stochastic environments

A. Modirshanechi, W.-H. Lin, H.A. Xu, M. Herzog, and W. Gerstner
Published in Proc. Natl. Acad. Sci. (USA), 122:e2502193122 in 2025

High-performance deep spiking neural networks with 0.3 spikes per neuron

A. Stanojevic, S. Wozniak, G. Bellec, G. Cherubini, A. Pantazi, and W. Gerstner
Published in Nature Communications, 168: 74-88 in 2024

Local plasticity rules can learn deep representations using self-supervised contrastive predictions

B. Illing, J. Ventura, G. Bellec, and W. Gerstner
Published in 35th Conference on Neural Information Processing Systems in 2021

Diverse synaptic plasticity mechanisms orchestrated to form and retrieve memories in spiking neural networks

F. Zenke, E.J. Agnes, and W. Gerstner
Published in Nature Communications, 6:6922 in 2015

Inhibitory Plasticity Balances Excitation and Inhibition in Sensory Pathways and Memory Networks

T. Vogels, H. Sprekeler, F. Zenke, C. Clopath and W. Gerstner
Published in Science 334: 1569-1573 in 2011

Connectivity reflects coding: a model of voltage-based STDP with homeostasis

C. Clopath, L. Busing, E. Vasilaki and W. Gerstner
Published in Nature Neuroscience, 13: 344-352 in 2010

Research

Current Research Fields

Spiking Neurons
Biologically Plausaible Learning Rules
Loss Landsacpe of Neural Networks

Teaching & PhD

Computational Neuroscience: Neuronal Dynamics.

For Master students in NeuroX, Physics, Life Science Engineering, Computer Science, Mathematics/Computational Science.

This course covers mathematical models of neurons and neuronal networks in the context of biology and establish links to models of cognition. The focus is on brain dynamics approximated by deterministic or stochastic differential equations. 
Hodgkin-Huxeley model, Hopfield model, Decision model,  stochastic leaky integrate-and-fire model, spike response model, phase plane analysis, Poisson process, Renewal process, mean-field methods for dynamics.

Brain-Style Learning in Neural Networks.

Master students in Computer Science and NeuroX.

 Biological brains show powerful learning without BackProp, how? By a smart combination of Reinforcement Learning and Self-supervised learning with local learning rules at the connections (synapses). 

 - Why BackProp is biologically not plausible. 
- Two-factor and three-factor rules  in biology (synaptic plasticity) and neuromorphic hardware (
- Three-factor rules for reward-based learning 
- Reinforcement Learning in the brain
- Learning of efficient representations with biologically plausible learning rules



PhD Students

Becker Sophia, Gross Alisa, Delrocq Ariane, Nguyen Tâm Johan, Gruaz Lucas Louis, Pezon Louis Henry, Wu Zihan, Wang Shuqi, Martinelli Flavio, Smeets Kasper

Past EPFL PhD Students

Mona Spiridon Paltani, Fabrizio Smeraldi, Angelo Arleo, Perry Moerland, Alix Herrmann Scheurer, Felix Gers, Pierre-Edouard Sottas, Silvio Borer, Thomas Strösslin, Julien Mayor, Ricardo Andres Chavarriaga Lozano, Renaud Jolivet, Jean-Pascal Théodor Pfister, Denis Sheynikhovich, Brice Bathellier, Laurent Badel, Claudia Clopath, Nicolas Marcille, Richard Naud, Christian Tomm, Nicolas Frémaux, Guillaume Hennequin, Skander Mensi, Christian Antonio Pozzorini, Lorric Ziegler, Friedemann Zenke, Felipe Gerhard, Mohammadjavad Faraji, Carlos Stein Naves de Brito, Alex Seeholzer, Hesam Setareh, Marco Philipp Lehmann, Dane Sterling Corneil, Samuel Pavio Muscinelli, Olivia Gozel, Florian François Colombo, Vasiliki Liakoni, Chiara Gastaldi, Bernd Albert Illing, Valentin Schmutz, Martin Louis Lucien Rémy Barry, Ana Stanojevic, Alireza Modirshanechi

Past EPFL PhD Students as codirector

Laurence Meylan, Gediminas Luksys, Marius Kleiner, Danilo Jimenez Rezende, Georgios Iatropoulos, Berfin Simsek, Sourmpis Christos

Courses

Computational neurosciences: neuronal dynamics

NX-465

In this course we study mathematical models of neurons and neuronal networks in the context of biology and establish links to models of cognition. The focus is on brain dynamics approximated by deterministic or stochastic differential equations.

Learning in neural networks

CS-479

Full title: "Brain-style learning in Neural Networks: Learning algorithms of the brain". Biological brains show powerful learning without BackProp, how? By a smart combination of Reinforcement Learning and Self-supervised learning with local learning rules at the connections (synapses).