← All Papers · Machine Learning

Emergent Capabilities as Universal Latent Modes

Tamás Nagy, Ph.D. Updated 2026-03-20 Short Draft Machine Learning
Unreviewed draft. This paper has not been human-reviewed. Mathematical claims may be unverified. Use with appropriate caution.
Download PDF View in Graph BibTeX

Abstract

Emergent capabilities in large language models — abilities that appear suddenly as model scale increases — remain poorly understood. We propose that emergence corresponds to the activation of new universal Latent modes: directions in the model's representation space that are shared across all domains. Using the Latent of Latents framework (Nagy 2026h), we decompose a model's knowledge into domain-specific variation (captured by centered meta-axes) and universal variation (captured by the non-centered mean component). Each new universal mode represents a cross-domain capability — summarization, reasoning, analogy — that the model has learned to apply uniformly.

Preliminary evidence from GPT-2 Small (1 universal mode, 99.6% of universal variance) and TinyLlama 1.1B (2–3 universal modes, 67% + 3.5% + 0.8%) shows that larger models develop more universal modes, consistent with the emergence of new cross-domain capabilities at scale. We propose a tri-graded Latent algebra \(\Lambda^{(i,j,k)}\) where \(i\) = within-domain, \(j\) = across-domain, \(k\) = universal depth, and predict that the number of universal modes at threshold \(\varepsilon\) scales as \(K_\varepsilon \sim \log \log P\) where \(P\) is parameter count.

Length
3,116 words
Claims
1 theorems
Status
Draft (Outline + Preliminary Data)
Target
ICML 2027 / Nature Machine Intelligence

Connects To

The Latent of Latents: Hierarchical Finite Representations o... The Latent: Finite Sufficient Representations of Smooth Syst... Machine-Verified Spectral Intelligence: From Data Geometry t... Universal Foundations: A Verified Library of Core Mathematic...

Browse all Machine Learning papers →