The most complex model we actually understand

Welch Labs
Dec 20, 2025
14 notes
14 Notes in this Video

Modular Arithmetic as Neural Network Learning Task

ModularArithmetic TrainingData ClockArithmetic BoundedOperations
01:11

The OpenAI research team studying grokking chose modular arithmetic as their experimental task, creating datasets by systematically varying operands and computing remainders after division by a modulus.

One-Hot Encoding: How Models Perceive Tokens

OneHotEncoding TokenRepresentation SparseVectors InputEncoding
01:59

Neural networks receive input through one-hot encoding, a sparse representation where each token activates exactly one position in a vector while all others remain zero.

Grokking: Delayed Generalization in Neural Networks

Grokking Generalization TrainingDynamics DelayedLearning
03:05

Discovered accidentally by an OpenAI research team in 2021 when a researcher left a model training during vacation, the term “grokking” comes from Robert Heinlein’s 1961 novel describing profound understanding that merges with the knower.

Embedding Matrix Transforms Sparse Inputs to Dense Representations

EmbeddingMatrix DenseVectors LearnedWeights RepresentationLearning
04:03

The embedding matrix multiplies sparse one-hot input vectors to produce dense embedding vectors, serving as the first learned transformation in the network.

Network Learns Fourier Trigonometric Solution to Modular Addition

FourierAnalysis TrigonometricFunctions PeriodicSolutions FrequencyDomain
05:46

Neil Nandanda’s team discovered through discrete Fourier analysis that the network learns to represent inputs and operations using sine and cosine functions at specific frequencies.

Sparse Linear Probes Extract Hidden Trigonometric Signals

LinearProbes InterpretabilityMethods SignalExtraction WeightedSums
06:26

Researchers analyzing the grokking model use sparse linear probes as an interpretability technique to extract clean trigonometric signals from noisy distributed embedding representations.

Analog Clocks Physically Implement Modular Arithmetic

ClockArithmetic CircularStructure PhysicalAnalogy PeriodicMotion
07:02

Analog clocks serve as everyday physical implementations of modular addition, with circular motion perfectly matching modulo arithmetic’s wraparound behavior.

Diagonal Symmetry in Activations Detects Input Sums

ActivationPatterns DiagonalStructure SumDetection GeometricLearning
09:14

Deeper multi-layer perceptron neurons develop diagonal wave patterns where activation crests align with specific input sum values, enabling sum detection through geometric structure.

Trigonometric Sum Identity Enables Neural Addition

TrigonometricIdentity AngleAddition MathematicalIdentities ProductToSum
09:52

The network discovers and effectively implements the trigonometric identity cos(x)cos(y) - sin(x)sin(y) = cos(x+y) to convert products of sines and cosines into the sum of angles itself.

Excluded Loss Reveals Hidden Progress During Grokking

ExcludedLoss TrainingMetrics FrequencyAnalysis HiddenProgress
11:16

Neil Nandanda and collaborators proposed the excluded loss metric as a clever diagnostic revealing internal structure formation invisible to standard accuracy and loss measurements.

Three Training Phases: Memorization, Structure Building, Cleanup

TrainingPhases MemorizationPhase CleanupPhase LearningDynamics
11:52

Nandanda and collaborators identified three distinct phases networks progress through when grokking: initial memorization, hidden structure building, and final cleanup.

Claude Haiku Represents Character Count on Six-Dimensional Manifold

ClaudeHaiku HighDimensionalRepresentation ManifoldLearning LineBreakMechanism
12:27

An Anthropic research team studying Claude 3.5 Haiku discovered clean geometric structure controlling the model’s line-break insertion behavior, representing a rare interpretability success in full-size models.

QK Twist: Attention Rotates Manifolds to Detect Distance

AttentionMechanism GeometricRotation QKTwist DistanceDetection
12:52

The Anthropic team discovered what they term a “QK twist” where attention blocks rotate helix-like geometric representations in six-dimensional space to implement distance comparison.

AI as Alien Intelligence: Summoning Ghosts, Not Building Minds

AIPhilosophy AlienIntelligence Anthropomorphism ConceptualFrameworks
13:31

AI researcher Andre Karpathy suggested that training large language models resembles summoning ghosts more than building animal intelligence, framing AI as fundamentally alien rather than human-like.