This episode currently has no reviews.
Submit ReviewCheck out my free video series about what's missing in AI and Neuroscience
Support the show to get full episodes and join the Discord community.
Matthew Larkum runs his lab at Humboldt University of Berlin, where his group studies how dendrites contribute to computations within and across layers of the neocortex. Since the late 1990s, Matthew has continued to uncover key properties of the way pyramidal neurons stretch across layers of the cortex, their dendrites receiving inputs from those different layers - and thus different brain areas. For example, layer 5 pyramidal neurons have a set of basal dendrites near the cell body that receives feedforward-like input, and a set of apical dendrites all the way up in layer 1 that receives feedback--like input. Depending on which set of dendrites is receiving input, or neither or both, the neuron's output functions in different modes- silent, regular spiking, or burst spiking. Matthew realized the different sets of dendritic inputs could signal different operations, often pairing feedforward sensory--like signals and feedback context-like signals. His research has shown this kind of coincidence detection is important for cognitive functions like perception, memory, learning, and even wakefulness. We discuss many of his ideas and research findings, why dendrites have long been neglected in favor of neuron cell bodies, the possibility of learning about computations by studying implementation-level phenomena, and much more.
0:00 - Intro 5:31 - Background: Dendrites 23:20 - Cortical neuron bodies vs. branches 25:47 - Theories of cortex 30:49 - Feedforward and feedback hierarchy 37:40 - Dendritic integration hypothesis 44:32 - DIT vs. other consciousness theories 51:30 - Mac Shine Q1 1:04:38 - Are dendrites conceptually useful? 1:09:15 - Insights from implementation level 1:24:44 - How detailed to model? 1:28:15 - Do action potentials cause consciousness? 1:40:33 - Mac Shine Q2
Check out my free video series about what's missing in AI and Neuroscience
Support the show to get full episodes and join the Discord community.
Matthew Larkum runs his lab at Humboldt University of Berlin, where his group studies how dendrites contribute to computations within and across layers of the neocortex. Since the late 1990s, Matthew has continued to uncover key properties of the way pyramidal neurons stretch across layers of the cortex, their dendrites receiving inputs from those different layers - and thus different brain areas. For example, layer 5 pyramidal neurons have a set of basal dendrites near the cell body that receives feedforward-like input, and a set of apical dendrites all the way up in layer 1 that receives feedback--like input. Depending on which set of dendrites is receiving input, or neither or both, the neuron's output functions in different modes- silent, regular spiking, or burst spiking. Matthew realized the different sets of dendritic inputs could signal different operations, often pairing feedforward sensory--like signals and feedback context-like signals. His research has shown this kind of coincidence detection is important for cognitive functions like perception, memory, learning, and even wakefulness. We discuss many of his ideas and research findings, why dendrites have long been neglected in favor of neuron cell bodies, the possibility of learning about computations by studying implementation-level phenomena, and much more.
0:00 - Intro 5:31 - Background: Dendrites 23:20 - Cortical neuron bodies vs. branches 25:47 - Theories of cortex 30:49 - Feedforward and feedback hierarchy 37:40 - Dendritic integration hypothesis 44:32 - DIT vs. other consciousness theories 51:30 - Mac Shine Q1 1:04:38 - Are dendrites conceptually useful? 1:09:15 - Insights from implementation level 1:24:44 - How detailed to model? 1:28:15 - Do action potentials cause consciousness? 1:40:33 - Mac Shine Q2
This episode currently has no reviews.
Submit ReviewThis episode could use a review! Have anything to say about it? Share your thoughts using the button below.
Submit Review