Researchers at EPFL’s Blue Brain Project have discovered that dendrites, the neuron’s tree-like receptors, work together — dynamically and depending on the workload — for learning. The findings further our understanding of how we think and may inspire new algorithms for artificial intelligence.
The analysis was performed using cells from the Blue Brain’s virtual rodent cortex. The researchers expect other types of neurons — non-cortical or human — to operate in the same way.
These results show that when a neuron receives input, the branches of the elaborate tree-like receptors extending from the neuron, known as dendrites, functionally work together in a way that is adjusted to the complexity of the input.
How Do Neurons Process Information?
The strength of a synapse determines how strongly a neuron feels an electric signal coming from other neurons, and the act of learning changes this strength.
By analyzing the “connectivity matrix” that determines how these synapses communicate with each other, the algorithm establishes when and where synapses group into independent learning units from the structural and electrical properties of dendrites. In other words, the new algorithm determines how the dendrites of neurons functionally break up into separate computing units and finds that they work together dynamically, depending on the workload, to process information.
The researchers liken their results to the functioning of computing technology already implemented today.
Like Cloud Computing
This newly observed dendritic functionality acts like parallel computing units, meaning that a neuron is able to process different aspects of the input in parallel, like supercomputers.
Each of the parallel computing units can independently learn to adjust its output, much like the nodes in deep learning networks used in artificial intelligence (AI) models today. Comparable to cloud computing, a neuron dynamically breaks up into the number of separate computing units demanded by the workload of the input.
“In the Blue Brain Project, this mathematical approach helps to ascertain functionally relevant clusters of neuronal input which are inputs that feed into the same parallel processing unit. This then enables us to determine the level of complexity at which to model cortical networks as we digitally reconstruct and simulate the brain,”
explains Marc-Oliver Gewaltig, Section Manager in Blue Brain’s Simulation Neuroscience Division.
Compartmentalization Is Dynamic
Additionally, the research reveals how these parallel processing units influence learning, i.e. the change in connection strength between different neurons. The way a neuron learns depends on the number and location of parallel processors, which in turn depend on the signals arriving from other neurons.
For instance, certain synapses that do not learn independently when the neuron’s input level is low, start to learn independently when the input levels are higher.
To date, traditional learning algorithms (such as those currently used in A.I. applications) assume that neurons are static units that merely integrate and re-scale incoming signals. By contrast, the results show that the number and size of the independent subunits can be controlled by balanced input or shunting inhibition.
The researchers propose that this temporary control of the compartmentalization constitutes a powerful mechanism for the branch-specific learning of input features.
“The method finds that in many brain states, neurons have far fewer parallel processors than expected from dendritic branch patterns. Thus, many synapses appear to be in ‘grey zones’ where they do not belong to any processing unit. However, in the brain, neurons receive varying levels of background input and our results show that the number of parallel processors varies with the level of background input, indicating that the same neuron might have different computational roles in different brain states,”
explains lead scientist and first author Willem Wybo.
Willem A.M. Wybo, Benjamin Torben-Nielsen, Thomas Nevian, Marc-Oliver Gewaltig
Electrical Compartmentalization in Neurons
Cell Reports (2019). DOI: 10.1016/j.celrep.2019.01.074
Image: BBP / EPFL 2019