Gaute Einevoll public
[search 0]
More
Download the App!
show episodes
 
Loading …
show series
 
Numerous neuron models have been made, but most of them are "single-purpose" in that they are made to address a single scientific question. In contrast, multipurpose neuron models are made to be used to address many scientific questions. In 2011, the guest published a multipurpose rodent pyramidal-cell model which has been actively used by the comm…
  continue reading
 
With modern electrical and optical measurement techniques, we can now measure neural activity in hundreds or thousands of neurons simultaneously. This allows for the investigation of population codes, that is, of how groups of neurons together encode information. In 2019 today’s guest published a seminal paper with collaborators at UCL in London wh…
  continue reading
 
The observed variety of dendritic structures in the brains is striking. Why are they so different, and what determine the branching patterns? Following the dictum “if you understand it, you can build it”, the lab of the guest builds dendritic structures in a computer and explore the underlying principles. Two key principles seem to be to minimize (…
  continue reading
 
The term “foundation model” refers to machine learning models that are trained on vast datasets and can be applied to a wide range of situations. The large language model GPT-4 is an example. The group of the guest has recently presented a foundation model for optophysiological responses in mouse visual cortex trained on recordings from 135.000 neu…
  continue reading
 
A holy grail of the multiscale approach for physical brain modelling is to link the different scales from molecules, via cells and local neural networks, up to whole-brain models. The goal of the Virtual Brain Twin project, lead by today’s guest, is to use personalized human whole-brain models to aid clinicians in treating brain ailments. The podca…
  continue reading
 
In 1982 John Hopfield published the paper "Neural networks and physical systems with emergent collective computational abilities" describing a simple network model functioning as an associative and content-addressable memory. The paper started a new subfield in computational neuroscience and led to the influx of numerous theoretical scientists, in …
  continue reading
 
The leading theory for learning and memorization in the brain is that learning is provided by synaptic learning rules and memories stored in synaptic weights between neurons. But this is for long-term memory. What about short-term, or working, memory where objects are kept in memory for only a few seconds? The traditional theory held that here the …
  continue reading
 
In September Paul Middlebrooks, the producer of the podcast BrainInspired, and I were both on a neuro-AI workshop on a coast liner cruising the Norwegian fjords. We decided to make two joint podcasts with some of the participants where we discuss the role of AI in neuroscience. In this second part we discuss the topic with Cristina Savin and Tim Vo…
  continue reading
 
In September Paul Middlebrooks, the producer of the podcast BrainInspired, and I were both on a neuro-AI workshop on a coast liner cruising the Norwegian fjords. We decided to make two joint podcasts with some of the participants where we discuss the role of AI in neuroscience. In this first part we talk with Mikkel Lepperod, the main organizer abo…
  continue reading
 
Most of what we have learned about the functioning of the living brain has come from extracellular electrical recordings, like the measurement of spikes, LFP, ECoG and EEG signals. And most analysis of these recordings has been statistical, looking for correlations between the recorded signals and what the animal/human is doing or being exposed to.…
  continue reading
 
The most prominent visual characteristic of neurons is their dendrites. Even more than 100 years after their first observation by Cajal, their function is not fully understood. Biophysical modeling based on cable theory is a key research tool for exploring putative functions, and today’s guest is one the leading researchers in this field. We talk a…
  continue reading
 
The greatest mystery of all is why a group of atoms, like the ones constituting me, can feel anything. The mind-brain problem has puzzled philosophers for millennia. Thanks to pioneers like Christof Koch, consciousness studies have recently become a legitimate field of scientific inquiry. In this vintage episode, recorded in February 2021, we discu…
  continue reading
 
Computational neuroscientists use many software tools, and NEURON has become the leading tool for biophysical modeling of neurons and neural network. Today’s guest has been the leading developer of NEURON since the infancy almost 50 years ago. We talk about how the tool got started and the development up until today’s modern version of the software…
  continue reading
 
The idea that memories are stored in molecules was popular in the middle of the 20th century. However, since the discovery of long-term potentiation (LTP) in the 1970s, the dominant view has been that our memories are stored in synapses, that is, in the connections between neurons. Today, there are signs that the interest in molecular memory is ret…
  continue reading
 
Is quantum physics important in determining how living systems, including brains, work? Today's guest is a professor of molecular genetics at the University of Surrey in England and explores this question in the book “Life at the edge: The coming of age of quantum biology”. In this “vintage” episode, recorded in late 2019, we talk about how quantum…
  continue reading
 
Most computational neuroscientists investigate electric dynamics in neurons or neural networks, but there is also computations going on inside neurons. Here the key dynamical variables are concentrations of numerous different molecules, and the signaling is typically done in cascades of chemical reactions, called signaling pathways. Today’s guest i…
  continue reading
 
Today’s AI is largely based on supervised learning of neural networks using the backpropagation-of-error synaptic learning rule. This learning rule relies on differentiation of continuous activation functions and is thus not directly applicable to spiking neurons. Today’s guest has developed the algorithm SuperSpike to address the problem. He has a…
  continue reading
 
Over the last ten years or so, the MindScope project at the Allen Institute in Seattle has pursued an industrylab-like approach to study the mouse visual cortex in unprecedented detail using electrophysiology, optophysiology, optical imaging and electron microscopy. Together with collaborators at Allen, today’s guest has worked to integrate of thes…
  continue reading
 
Today’s guest is a pioneer both in the fields of computational neuroscience and artificial intelligence (AI) and has had a front seat during their development. His many contributions include, for example, the invention of the Boltzmann machine with Ackley and Hinton in the mid 1980s. In this “vintage” episode recorded in late 2019 he describes the …
  continue reading
 
Today’s guest has argued that the present dominant way of doing systems neuroscience in mammals (large-scale electric or optical recordings of neural activity combined with data analysis) will be inadequate for understanding how their brain works. Instead, he proposes to focus on the simple roundworm C.elegans with only 302 neurons and try to rever…
  continue reading
 
Over the last decade topological analysis has been established as a new tool for analysis of spiking data. Today’s guest has been a pioneer in adapting this mathematical technique for use in our field and explains concepts and example applications. We also also talk about so-called threshold-linear network model, a generalization of Hopfield networ…
  continue reading
 
Not all interesting network activity occurs in cortex. Networks in the spinal cord, the long thin tubular structure extending downwards from the neck, is responsible for setting up rhythmic motor activity needed for moving around. How do these so-called central pattern generators work? Today’s guest has, together with colleagues in Copenhagen, deve…
  continue reading
 
We know a lot about of how neurons in the primary visual cortex (V1) of mammals respond to visual stimuli. But how does the vast information contained in the spiking of millions of neurons in V1 give rise to our visual percepts? The guest’s theory is that V1 acts as a “saliency detector” directing the gaze to the most important object in the visual…
  continue reading
 
A key goal of computational neuroscience is to build mathematical models linking single-neuron activity to systems-level activity. The guest has taken some bold steps in this direction by developing and exploring a multi-area model for the macaque visual cortex, and later also a model for the human cortex, using millions of simplified spiking neuro…
  continue reading
 
It is widely thought that spikes (action potentials) are the main carrier of information in the brain. But what is the neural code, that is, what aspects of the spike trains carry the information? The detailed temporal structure or maybe only the average firing rate? And is there information in the correlation between spike trains in populations of…
  continue reading
 
Starting from the pioneering work of Hodgkin, Huxley and Rall in the 1950s and 60s, we have a well-founded biophysics-based mathematical understanding of how neurons integrate signals from other neurons and generate action potentials. Today’s guest wrote the classic book “Biophysics of Computation” on the subject in 1998. We discuss its contents, w…
  continue reading
 
The book “Models of the Mind” published in 2021 gives an excellent popular account of the history and questions of interest in theoretical neuroscience. I could think of no other person more suitable to invite for the inaugural episode of the podcast than its author Grace Lindsay. In the podcast we discuss highlights from the book as well as recent…
  continue reading
 
Loading …

Quick Reference Guide

Listen to this show while you explore
Play