I am a researcher in computational neuroscience at the Group for Neural Theory of Ecole Normale Supérieure. I use tools from statistical physics, machine learning and data analysis to understand how cognition emerges from the dynamics of networks of neurons.
Theory of attractor neural networks
Attractor neural networks describe a class of recurrent networks, the dynamics of which generates trajectories that are attracted towards specific sets in state-space. Many theoretical tools have been developed over the last 30 years to study the properties of these trajectories and relate them to neural computations, such as those involved in working memory or spatial navigation.
Analysis of biological
Experimental techniques developing at a fast pace allow to record more and more neurons at increasing resolutions in behaving animals. This provides a unique opportunity to advance our understanding of brain functions, typically by i) mapping animal behavior onto recorded neural dynamics, and ii) accounting for this dynamics from the anatomy and physiology of the underlying neural substrate.
Analysis of artificial
Recent advances in machine learning have allowed to train neural networks to perform a variety of tasks. These functional networks, which are the outcome of a training algorithm, are often called black-boxes and a major scientific challenge is to develop tools to open them. This is for instance crucial in applications of machine learning, where interpretability of the behavior of a trained network can be required to gain insight into underlying mechanisms.