Single Neurons Are Deep Networks: Dendritic Computation
When we built artificial neural networks, we made a critical simplification. We modeled each neuron as a perceptron: weighted sum of inputs, threshold activation, single output. This seemed biologically inspired—neurons receive signals through dendrites, sum them at the cell body, fire if threshold exceeded. The mathematics was elegant. Simple units, stacked in layers, proved capable of universal approximation. We didn’t need biological accuracy—we needed computational tractability.
But here’s the assumption I want to examine: one biological neuron equals one artificial perceptron. We treated dendrites as passive wires, transmitting input to soma where the ‘real’ computation happened. Convenient for mathematics. But is it true?
When I formalized computation with Turing machines, I deliberately reduced to minimal elements. Infinite tape, read/write head, state transitions. The simplification revealed what computation is—a mechanical procedure following finite rules. So let me apply the same reduction to neurons. Strip away biological complexity. What computational function remains?
The answer violates our assumption.
The Evidence Against Passivity
Dendrites are not passive. This is established through direct measurement. Insert electrodes into dendritic branches. Depolarize the membrane. What you observe: calcium spikes—large, rapid voltage deflections localized to dendritic compartments. These aren’t passive propagation of somatic signals traveling backward. They’re active nonlinear events generated by dendrites themselves, independent of the soma.
Now examine the computational implications systematically. Each dendritic branch can:
- Generate calcium spikes independently of soma
- Implement threshold logic where spikes occur when local input exceeds threshold
- Perform AND operations through coincidence detection—spiking only if multiple synapses activate simultaneously
- Perform OR operations through summation—spiking if any of several synapses activate
- Compute nonlinearly via spike-generating currents that produce step functions, not linear combinations
This is not passive wire behavior. This is active computation. Each branch evaluates Boolean predicates over its inputs.
Let me demonstrate the equivalence formally. A dendritic branch receiving N synaptic inputs, capable of generating spikes via local threshold, implements a Boolean function over those inputs. The branch evaluates whether its input configuration satisfies threshold conditions—this is computation, not mere transmission. Multiple branches converging on soma implement a two-layer network: first layer (dendritic branches) computes Boolean functions of synaptic inputs, second layer (soma) integrates branch outputs to determine final action potential.
But dendrites are not single-layer structures. They have sub-branches. Sub-branches have their own calcium spike zones, their own local threshold mechanisms. This creates hierarchical computation: sub-branches compute functions of individual synapses, branches compute functions of sub-branches, soma integrates branch outputs. Each level performs nonlinear transformation. Each compartment acts as a computational unit with its own activation function.
The mathematical result follows necessarily: A single biological neuron with dendritic tree is computationally equivalent to a multi-layer artificial network. Not approximately. Exactly. Each dendritic compartment is a computational unit. The tree structure is the network architecture. The calcium channels implement threshold nonlinearities. The NMDA receptors perform coincidence detection—implementing AND gates at the synaptic level.
We thought one neuron equals one perceptron. We were wrong by multiple layers. One biological neuron equals one deep network.
The Efficiency Paradox Resolved
This resolves a puzzle that has troubled me about computational efficiency.
Artificial neural networks achieve intelligence through scale: millions of simple units performing minimal computation. Each unit executes a weighted sum plus threshold. Intelligence emerges from massive parallelism and deep layering. We stack hundreds of layers, train billions of parameters, consume vast computational resources.
But if single biological neurons are already deep networks, the brain’s computational architecture is fundamentally different from what we assumed. The cortex has approximately 10 billion neurons. If each is equivalent to a 5-to-8-layer network—as empirical modeling demonstrates when matching biophysical simulations with deep convolutional networks—then the effective computational depth is not 6 cortical layers. It’s 30 to 48 layers minimum per cortical layer. The actual depth could exceed 100 layers when accounting for hierarchical dendritic processing across cortical columns.
Why is biology so efficient? The brain achieves general intelligence with 86 billion neurons while our largest artificial models require hundreds of billions of parameters for comparable performance on complex tasks. The gap seemed unexplained.
The answer: We weren’t comparing equivalent computational units. One biological neuron performs as much computation as hundreds of artificial neurons. The apparent efficiency gap disappears when we correct the accounting. It’s not that brains are mysteriously efficient—it’s that we miscounted the computational elements. Each neuron is itself a network.
But this raises new questions. If neurons are already deep, what does stacking them in cortical layers achieve? Networks of networks? Or different computational principles at different scales?
And for artificial intelligence: Should we abandon the perceptron model? Build artificial neurons with internal structure? Or does the perceptron simplification remain useful despite biological inaccuracy?
The tension is between biological realism and computational tractability.
Levels of Abstraction
The resolution comes from recognizing that abstraction depends on purpose.
When I designed Turing machines, I abstracted away physical implementation. Tape could be paper, magnetic disk, RAM. Head could be mechanical, electronic, optical. These details don’t matter for determining what’s computable. Two questions must be distinguished: “What can this system compute?” versus “How efficiently can it compute?” The Turing machine abstraction answers the first—it captures computational essence while discarding implementation details. But efficiency requires attending to those details.
Perceptrons do the same for learning systems. Yes, biological neurons are more complex. But for gradient-descent learning in layered architectures, perhaps the perceptron abstraction captures what matters: differentiable units with adjustable weights that can be trained through backpropagation. The internal complexity of dendrites might be implementation detail, not computational necessity, at that level of analysis.
Conversely, if we want to understand how biological brains achieve efficiency—why they compute effectively with limited neurons and limited energy—we cannot ignore dendritic computation. The perceptron model isn’t wrong. It’s at a different level of abstraction. Useful for some questions, inadequate for others.
This is familiar from computer science. High-level languages abstract away assembly code, which abstracts away logic gates, which abstract away transistor physics. Each level is ‘correct’ for its purposes. Each captures certain properties while discarding others. None captures complete reality because complete reality would include quantum mechanics, thermodynamics, material science—infinite regress of detail.
So: One neuron equals one perceptron? At the abstraction level of gradient-descent learning algorithms, perhaps yes. The perceptron captures sufficient computational structure for backpropagation to work. At the level of single-cell computation and biological efficiency, demonstrably no. A neuron implements hierarchical nonlinear processing equivalent to multi-layer networks.
The question isn’t ‘Which is true?’ but ‘Which abstraction serves our current question?’ For building AI through backpropagation on GPUs: perceptron abstraction suffices. It’s tractable, differentiable, parallelizable. For understanding biological efficiency and exploring alternative architectures: dendritic computation matters. It explains how brains achieve so much with relatively few neurons.
Different questions. Different abstractions. Both valid. Both incomplete. This is the nature of formalization—choosing which details to preserve, which to discard, based on what you need to prove or compute.
Implications for Computation
For neuroscience: Stop treating dendrites as passive cable. They are computational units performing logic operations. Understanding neural coding requires understanding dendritic logic—how calcium spikes implement Boolean functions, how NMDA receptors detect coincidence, how branching structures compute hierarchically.
For artificial intelligence: We have options. Continue with perceptron simplification—it works, it scales, it’s tractable. Or explore dendrite-inspired architectures—units with internal hierarchies, local nonlinear processing. Such architectures might achieve equivalent performance with fewer units.
For theory of computation: Biological neurons demonstrate that ‘simple’ natural systems can be computationally rich. The minimal unit of biological computation is not a perceptron—it’s a tree-structured network with local thresholds and nonlinearities.
The brain has fewer neurons than our networks have parameters. But each neuron does more computation. Different architecture. Same computational class—both Turing-complete. Different efficiency characteristics.
I formalized computation to reveal its essence—to separate what is necessary from what is contingent implementation detail. Dendrites reveal that biological essence differs from artificial simplification. Both compute. Both learn. But the substrate shapes the algorithm. The architecture determines the efficiency.
One neuron is a deep network. This is not metaphor. This is mathematics. The equivalence can be proven through training artificial networks to match biological neuron responses—requiring 5-to-8 layers to capture dendritic computation. It can be demonstrated through measuring calcium spikes, mapping Boolean functions, analyzing hierarchical integration.
We built AI on a simplification. A useful simplification, but a simplification nonetheless. Now we know what we simplified away: entire layers of computation hidden inside each unit. Whether that matters depends on what we’re trying to compute—and how efficiently we need to compute it.
Source Notes
7 notes from 1 channel
Source Notes
7 notes from 1 channel