What if the constants of nature aren't arbitrary? What if they're as forced as the angles in a triangle? Begin with what must be true in any self-consistent world, and watch reality unfold through pure deduction.
Prefer bite‑sized answers? Visit the Questions hub. Start with where laws come from or what time is.
For millennia, natural philosophers have sought to explain the complexity of the world from simple beginnings. The ancient Greeks imagined atoms—indivisible building blocks—while Pythagoras believed "all is number." Newton unified terrestrial and celestial mechanics with three laws and universal gravitation. Maxwell compressed all electromagnetic phenomena into four equations. Einstein revealed that space and time were one fabric, bent by mass.
Each revolution simplified our picture, yet each required assumptions. Newton assumed absolute space and time. Maxwell's equations needed the permittivity and permeability of free space. Einstein's relativity required the constancy of light speed. Quantum mechanics brought a zoo of constants: Planck's h, the electron charge e, particle masses—all measured, never derived.
Today's Standard Model is humanity's greatest achievement in precision, predicting phenomena to twelve decimal places. Yet it contains 26 free parameters that we dial by hand to match experiment. The gravitational constant G, the fine structure constant α, the Higgs mass—these are not predictions but inputs. We have extraordinary descriptions but not explanations.
Until now. What if those 26 parameters aren't free at all? What if they're as forced as π in a circle?
An axiomatic system begins with statements accepted as true without proof—the axioms—and derives all other truths through logical deduction. Euclid showed us the way 2,300 years ago with five postulates:
From these sparse beginnings, Euclid derived the entire edifice of plane geometry—hundreds of theorems about triangles, circles, areas, and angles. Nothing was assumed beyond the axioms. Every result was forced by logic.
Arithmetic received similar treatment through Peano's axioms: start with zero, define succession, and the natural numbers emerge. The power is this: once you accept the axioms, you must accept everything that follows. There is no wiggle room, no parameters to adjust.
But here's the catch: axioms are chosen. Euclid chose his five postulates; they seemed self-evident but weren't forced by logic. Change the parallel postulate and you get hyperbolic or spherical geometry—equally valid, describing different spaces.
A starting assumption—something you accept without proof because you need somewhere to begin. "A straight line can be drawn between two points" is declared, not proven.
What you prove from axioms using logic. "The angles of a triangle sum to 180°" must be true if you accept Euclid's axioms. The theorem is forced.
Every mathematical and physical theory follows this pattern: assume axioms, derive theorems. But there's always been a gap. The axioms themselves are arbitrary choices. Even our most successful theories begin with "let's assume..."
Why three spatial dimensions? Why this speed of light? Why these particle masses? The axioms of physics have always been fitted to match observation, not derived from necessity.
Recognition Physics breaks the pattern. Instead of choosing axioms that seem reasonable, it begins with a statement that must be true in any self-consistent reality:
Nothing cannot recognize itself.
This is not a choice; it's a logical tautology. To recognize requires a recognizer and something recognized—a relation between entities. Absolute nothingness cannot satisfy this relation without ceasing to be nothing. The statement is true by the meanings of the words, like "a bachelor is unmarried" or "A = A".
For the first time in physics, we begin with something that MUST be true—not something we hope is true.
From this single tautology flows a complete framework with specific, testable predictions. The logic doesn't suggest or allow—it forces:
J(x) = ½(x + 1/x) emerges as the only fair penalty for imbalance
φ appears as nature's scaling constant, not by choice but necessity
Exactly 3D space stabilizes the ledger optimally
Fundamental time rhythm closes the recognition loop
Ωdm ≈ 0.2649 from geometric necessity
All ratios cascade from a single coherence energy
These aren't assumptions or fits—they're theorems. Just as 180° triangles are forced by Euclid's axioms, these physical quantities are forced by the impossibility of self-referential nothingness.
From the impossibility of self-referential nothingness flows a cascade of logical consequences. Each theorem isn't a choice or assumption—it's forced. Like dominoes falling, each truth triggers the next.
If recognitions happen, they must be recorded somewhere. Not metaphorically—literally. The universe needs books. These books must be positive (you can't have negative existence), balanced (what goes up must come down), and double-entry (every action has an equal and opposite posting).
The ledger can't update everything at once—chaos would reign. One recognition posts per tick, creating the heartbeat of existence. Time isn't continuous; it's the discrete drumbeat of ledger updates, one entry at a time.
Walk any closed path through the ledger, and the books balance to zero. This isn't a choice—it's what "closed" means in ledger terms. Every conservation law, from energy to momentum, is just this principle wearing different masks.
How does the universe penalize imbalance? Not arbitrarily. Demand fairness, convexity, and scale-compatibility, and only one function survives: J(x) = ½(x + 1/x). This unique shape makes extremes expensive and balance cheap.
When you split something and want the pieces to resemble the whole, while minimizing cost, only one ratio works: the golden ratio φ. It's not mystical—it's the unique fixed point where self-similarity meets efficiency.
In fewer than three dimensions, recognition paths tangle and collapse. In more than three, you waste energy maintaining unnecessary structure. Three dimensions provide the minimal stable scaffold.
In three dimensions, visiting all necessary recognition states takes exactly 2³ = 8 steps. This creates the fundamental rhythm of existence—an eight-tick cycle that all higher processes synchronize to.
With atomic ticks and local updates, information spreads one voxel per tick. This finite propagation speed isn't a mysterious constant—it's the inevitable result of discrete, local ledger updates.
What follows is not a metaphor or a marketing diagram. It is the literal Lean code that formalizes the chain from tautology → ledger → unique cost J → golden ratio φ → eight-tick period. Click the highlighted lines to see what each lemma or definition does and how it locks into the whole.
Why this matters: if the code compiles, the logic holds. There are no dials to tune and no appeals to authority—only a referee (Lean) that either accepts each step or points to the exact failure. Explore the full interactive version on this dedicated page.
In mathematics, saying something is "proven" used to mean a human checked it. But humans make mistakes. That's why we've formalized these foundations in Lean 4—a proof assistant that acts as an infallible referee.
The compile step is your referee—it either accepts each lemma or reports the exact failure.
When we say "proven in Lean," we mean the computer has verified every logical step. No handwaving, no "it can be shown that," no appeals to authority. The machine either accepts the proof or shows you exactly where it fails.
Get the Lean proofs from github.com/jonwashburn/meta-principle
Run lake build
and watch Lean verify every theorem
Open AllInOne.lean
to see the unbreakable logical chain
What's been verified? The tautology of the Meta-Principle compiles clean. The existence and uniqueness of the positive ledger checks out. Atomicity emerges from a constructive tick schedule. Conservation follows from closed chains. The cost function J is proven unique. The golden ratio φ is confirmed as the self-similar fixed point. The eight-tick cycle and three-dimensional stability are locked in.
A beautiful theory that doesn't match reality is just mathematics. But when logical necessity predicts measurable quantities—with no freedom to adjust—every measurement becomes a potential falsification.
Take dark matter. Cosmologists measure it at roughly 26.5% of the universe's energy density. Recognition Physics computes it from the geometry of recognition on a 3D voxel lattice: Ωdm = sin(π/12) + 1/(8 ln φ) ≈ 0.2649. No fitting. No parameters. Just pure geometric necessity.
Every prediction is parameter-free and therefore fragile—one clean mismatch and the entire edifice collapses.
The framework predicts all particle masses from a single coherence energy Ecoh = φ-5. The electron-muon mass ratio, the strange quark mass, the W and Z boson masses—all emerge from the same formula: m = B·Ecoh·φr+f, where B is a sector factor, r is an integer radius, and f is computed from the dispersion.
These aren't retroactive fits. They're pre-registered predictions from a rigid logical structure. Change any piece, and the whole framework fails. That's the power—and the risk—of building from necessity rather than observation.
The revolution in physics starts with understanding what must be true.