Signal and Noise: Information Theory Across Domains
The Fundamental Problem of Communication
The fundamental problem of communication is reproducing at one point a message selected at another. This statement, which opened my 1948 paper, appears deceptively simple. Yet it contains everything. A sender chooses from a set of possible messages. A receiver must reconstruct that choice. Between them lies a channel—and every channel introduces noise.
Consider Alice and Bob attempting to communicate across a wire. Every signal they transmit encounters interference: thermal fluctuations, electromagnetic disturbances, physical imperfections in the medium. The challenge is not merely sending a message but distinguishing intentional signals from environmental noise. This signal-to-noise problem defines the boundary conditions of all communication systems.
What surprised me, and what I believe remains underappreciated, is that this framework extends far beyond electrical engineering. The same mathematical structure appears wherever information flows: in the genetic instructions passed between generations, in the electrochemical signals neurons exchange, in every system that must transmit patterns reliably through an imperfect medium. Information, I came to understand, has universal properties independent of its physical substrate.
Bits, Channels, and the Entropy Bound
Let me define terms precisely. Information is the resolution of uncertainty. Before a message arrives, the receiver faces multiple possibilities. After receiving it, uncertainty decreases. The amount of this decrease, measured in bits, quantifies the information transmitted.
A channel’s capacity—the maximum rate at which information can be reliably transmitted—follows from a fundamental theorem: , where represents bandwidth, is signal power, and is noise power. This equation establishes an absolute limit. No encoding scheme, however clever, can exceed this capacity while maintaining reliability.
The implications are profound. First, noise does not merely corrupt messages—it imposes a ceiling on communication rate. Second, approaching this ceiling requires increasingly sophisticated encoding. Alice, transmitting dice roll outcomes, discovered that assigning shortest signals to most probable symbols and longer signals to rare outcomes optimizes throughput. This probability-based encoding approaches the theoretical limit by matching code length to information content.
Redundancy provides the mechanism for error correction. By adding controlled redundancy—extra bits that encode relationships between message elements—receivers can detect and correct errors introduced by noise. The overhead reduces raw transmission rate but increases reliable throughput. The optimal balance depends on channel characteristics: noisier channels require more redundancy.
DNA’s Redundant Alphabet
The genetic system, operating for billions of years before formal information theory existed, implements these same principles with remarkable sophistication.
DNA stores the accumulated learning of evolution as information—a repository of instructions encoded in a four-letter alphabet: adenine, thymine, guanine, cytosine. This is a discrete source: messages are selected from a finite symbol set, exactly as in communication theory. The genetic code translates this alphabet into proteins through triplet sequences called codons. Each three-nucleotide codon specifies one of twenty amino acids, forming the molecular machinery that creates organism structure and function.
Here is where the information-theoretic elegance becomes apparent. With four nucleotides and triplet codons, the system has possible codon combinations. Yet only twenty amino acids require encoding, plus start and stop signals. This creates massive redundancy: multiple codons specify the same amino acid. Leucine, for instance, is encoded by six different triplets.
This is not inefficiency—it is error correction. The redundancy in the genetic code means that single-nucleotide mutations often produce synonymous codons, leaving the protein unchanged. The system tolerates noise in the copying process because the encoding includes built-in fault tolerance. When thermal damage or copying errors introduce mutations, the degenerate code provides a buffer against functional disruption.
Some organisms push error correction further. Certain thermophilic archaea, inhabiting environments where heat continuously damages DNA, have developed DNA exchange mechanisms with neighboring cells. Rather than preventing thermal damage entirely—likely impossible at extreme temperatures—they accept ongoing errors and repair through homologous recombination. Cells share genetic material, using intact sequences from neighbors to reconstruct damaged regions. This is redundancy at the population level: the channel between generations maintains fidelity through distributed error correction.
Neural Population Codes
The brain faces its own version of the communication problem. Sensory neurons must encode environmental information into spike trains. Motor neurons must decode instructions into muscle contractions. At every stage, signals traverse noisy channels—synaptic transmission introduces variability, metabolic fluctuations corrupt signals, and the wetware itself operates near thermodynamic limits.
The fundamental encoding strategy appears in the McCulloch-Pitts model of neural computation: if input exceeds a threshold, the neuron fires; otherwise, it remains silent. This all-or-nothing response creates a discrete signaling system from continuous inputs. The binary spike provides noise immunity—the relevant information is whether a spike occurred, not its precise waveform. Amplitude variations, which would corrupt an analog signal, become irrelevant.
Yet single neurons are unreliable. Individual spike timing varies, thresholds fluctuate, and synaptic transmission fails stochastically. The brain’s solution parallels the thermophile strategy: population coding. Rather than relying on single neurons, the system distributes information across many parallel channels. A population of neurons collectively encodes a stimulus, with redundancy ensuring that individual failures do not corrupt the message.
Different neuron types implement different coding strategies optimized for different information types. Integrator neurons accumulate input over time, functioning as temporal averagers—they encode signal strength through cumulative effects, filtering high-frequency noise. Resonator neurons respond preferentially to specific frequencies, detecting temporal patterns in input timing. The brain deploys these computational types strategically, matching encoding schemes to the statistical structure of different signals.
Information Is Medium-Independent
The deepest insight from examining information flow across these domains is precisely what my original framework predicted: the mathematical structure remains invariant across substrates. Capacity limits apply whether you transmit bits, genes, or thoughts. Error correction through redundancy works whether implemented in parity bits, codon degeneracy, or neural population codes. The signal-to-noise ratio constrains all channels equally.
This universality is not metaphor—it is mathematics. Information theory provides the common language for analyzing communication in any domain. The same entropy calculations that optimize telegraph codes illuminate genetic evolution. The same channel capacity theorems that bound digital transmission rates constrain neural information processing.
The resolution of uncertainty, measured in bits, cares nothing for the physical medium carrying it. Information has properties that transcend implementation. This, perhaps, is the fundamental theorem worth remembering.
Source Notes
8 notes from 4 channels
Source Notes
8 notes from 4 channels