the algorithmic essence of superintelligence
from first principles: what are the minimum algorithms required for a self-proving, self-modeling, self-improving distributed intelligence?
sixteen components. one field. the system knows what it knows, proves it knows it, pays for what matters, grows itself optimally, and trusts no one.
0. the field: Goldilocks
everything starts here. $p = 2^{64} - 2^{32} + 1$. a prime with $2^{32}$ roots of unity — deep NTT support in a 64-bit word. four primitive operations over this field are sufficient for all computation in the system:
- fma (field multiply-accumulate): matrix ops, constraint evaluation, polynomial arithmetic
- ntt (number-theoretic transform): proof commitment, FHE polynomial multiply, quantum simulation
- p2r (Poseidon2 round): hashing, Fiat-Shamir, content addressing
- lut (lookup table): activation functions, S-box, FHE bootstrap, stark lookup
the same four primitives serve stark proofs, TFHE encryption, quantum simulation, and neural inference. not by coincidence — by algebraic structure. the field tower $\mathbb{F}_p \subset \mathbb{F}_{p^2} \subset \mathbb{F}_{p^3} \subset \mathbb{F}_{p^6}$ contains complex amplitudes (p²), recursive proof soundness (p³), and their composition (p⁶). one field. one chip (Goldilocks field processor). three pillars of the trinity: quantum security, privacy, AI.
see Goldilocks field for field properties. see nebu for field arithmetic implementation (1,526 LOC, 73 tests, DONE).
1. the hash: Hemera
Hemera: Poseidon2 sponge over Goldilocks. 24 rounds (8 full + 16 partial), 16-element state, rate 8 (56 bytes per absorption). self-bootstrapping: round constants generated by hashing "cyber" with zero-constant Hemera.
~736 stark constraints per permutation — but hemera is the trust anchor, not the bulk commitment engine. ~3 calls per execution: (1) domain separation wrapper for noun identity hemera(Lens.commit ‖ tag), (2) Fiat-Shamir seed, (3) Brakedown binding. lens (recursive Brakedown) handles bulk commitment. hemera wraps Lens commitments with domain tags. ~3 permutations per execution, not thousands.
one hash function, one security assumption (collision resistance), one implementation to audit.
see Hemera for specification. see hemera repo (5,084 LOC, 209 tests, PRODUCTION).
2. the object: cybergraph
a directed weighted graph $G = (P, E, N, w)$. particles $P$ are content-addressed by Hemera hash (32 bytes). cyberlinks $E$ are edges signed by neurons $N$ and weighted by stake $w$.
every particle is a complete semantic unit — 32-byte CID token vs 4-byte BPE fragment. unambiguous (hash = identity), universal (any content), no vocabulary freeze (graph grows).
the graph grows append-only (axiom A3). new particles and links accumulate. nothing is deleted. the cybergraph is the substrate — all dynamics, trust, intelligence operate on it.
3. the signal: atomic state change
a signal $s = (\nu, \vec\ell, \pi_\Delta, \sigma, \text{prev}, \text{mc}, \text{vdf}, \text{step})$:
- $\nu$: neuron identity (who)
- $\vec\ell$: batch of cyberlinks (what)
- $\pi_\Delta$: focus impulse — sparse local tri-kernel recomputation (the key field)
- $\sigma$: zheng proof covering cyberlinks + impulse + conviction (proof of everything)
- prev: hash of author's previous signal (hash chain)
- mc: Merkle clock (compact causal state)
- vdf: VDF proof (physical time since previous)
- step: monotonic counter (logical clock)
the impulse $\pi_\Delta$ is critical. the neuron does not just submit links — it computes the LOCAL effect of those links on the tri-kernel fixed point and includes the result. the zheng proof σ certifies this computation is correct. the network does not need to recompute φ* from scratch — it applies proven impulses incrementally.
this is the mechanism that makes reward ∝ Δπ work: the neuron proves its own Δπ inside the signal. the proof IS the claim. the claim IS the mining.
signals are patches. patches are signals. the same object described by patch theory (categorical morphism) and structural sync (verified broadcast unit). this is the cyb/fs/sync unification.
4. the dynamics: tri-kernel convergence
three operators on the graph simultaneously:
$$\phi^{(t+1)} = \text{norm}\left[\lambda_d \cdot \mathcal{D}(\phi^t) + \lambda_s \cdot \mathcal{S}(\phi^t) + \lambda_h \cdot \mathcal{H}_\tau(\phi^t)\right]$$
- $\mathcal{D}$ (diffusion): stake-weighted random walk. finds hubs
- $\mathcal{S}$ (springs): screened Laplacian, mean neighbor focus. finds equilibria
- $\mathcal{H}_\tau$ (heat): 2-hop smoothing at resolution τ. finds clusters
the collective focus theorem: the composite is contractive ($\kappa < 1$), unique fixed point $\phi^*$ exists, every initial distribution converges exponentially. $\phi^*$ IS focus — the consensus ranking.
two modes of computation:
-
incremental (per signal): each signal carries $\pi_\Delta$ — the neuron's local tri-kernel recomputation for its cyberlinks. the network accumulates proven impulses. φ updates continuously as signals arrive — no global recomputation per block
-
global (per epoch): full tri-kernel iteration from committed graph state. 23 iterations on bostrom, sub-second on GPU. validates accumulated impulses. with algebraic state commitments: fits in zheng circuit (1.42B constraints, 33% capacity) — provable
the impulse design means: a neuron contributing knowledge immediately shifts φ in its neighborhood. the shift propagates through subsequent signals. consensus emerges from accumulated local impulses, verified by periodic global proofs. the neuron proves its own Δπ inside the signal — the proof IS the mining.
5. the trust: five verification layers
| layer | discipline | mechanism | guarantee | cost |
|---|---|---|---|---|
| validity | computation | zheng proof | state transition correct | verify: ~5 μs |
| ordering | data structure | hash chain + VDF | operations carry own order | detect equivocation: O(1) |
| completeness | logic | lens / polynomial | nothing omitted | O(1) Lens opening |
| availability | probability | DAS + erasure coding | data physically exists | O(√n) samples |
| merge | algebra | CRDT / foculus | convergence deterministic | lattice union or π-convergence |
the composition: Verified Eventual Consistency (VEC). convergence guaranteed + completeness verifiable + availability verifiable. stronger than eventual consistency. a node proves it has converged, not assumes it.
conjectured minimum: remove any layer → a failure mode opens that the others cannot cover.
6. the proof: zheng
zheng: SuperSpartan IOP + recursive Brakedown lens + sumcheck, all over Goldilocks with Hemera.
nox trace → multilinear polynomial f(x₁,...,x_{n+4})
→ Brakedown commit (no Merkle — direct linear code)
→ SuperSpartan (CCS constraints via sumcheck)
→ proof π (~2 KiB)
prover: O(N) time, O(√N) memory (tensor compression). verifier: ~5 μs. recursive: the verifier IS a nox program — proofs compose. HyperNova folding: ~8K constraints → 30 field ops per recursive step (recursive Brakedown verifier is smaller than Merkle-based alternatives).
zheng: SuperSpartan IOP + recursive Brakedown lens (Goldilocks) + Binius lens (F₂). 32-64× cheaper for bitwise operations. cross-algebra boundary: ~766 F_p constraints.
see zheng for spec. see stark for the proof taxonomy.
7. the VM: nox
nox: 16 deterministic reduction patterns over binary trees of field elements. the computation substrate. every noun is a multilinear polynomial. axis = O(1) polynomial evaluation via Lens opening. identity = hemera(Lens.commit(noun) ‖ tag).
| layer | patterns | what they do |
|---|---|---|
| 1 | 0-15 | axis, quote, compose, cons, branch, add, sub, mul, inv, eq, lt, bit ops, hash |
| 2 | 16 (hint) | prover injects witness |
| 3 | 3 jet categories | 5 verifier jets + 8 binary jets + state jets (TRANSFER, INSERT, UPDATE patterns) |
the 16 patterns ARE the axiom. state operations, CCS jets, zheng itself — all derived from the 16 patterns.
algebra-polymorphic: nox<F, W, H> parameterized over field, word, hash. the same 16 patterns work over Goldilocks, F₂, F_{p²}. trident compiles 14 languages through nox.
focus metering: each pattern costs focus. insufficient focus → halt. the VM cannot loop forever — metering is structural, not a gas limit that can be circumvented.
8. the state: polynomial BBG
BBG: the authenticated state layer. polynomial commitments replace hash trees.
$$\text{BBG\_root} = H(\text{commit}(\text{BBG\_poly}) \| \text{commit}(A) \| \text{commit}(N))$$
BBG_poly covers 10 public dimensions (particles, axons_out, axons_in, neurons, locations, coins, cards, files, time, signals). A(x) is the commitment polynomial for membership (cyberlinks). N(x) is the nullifier polynomial for non-membership (spent). two independent private commitments alongside the public state polynomial. cross-index consistency: free (same polynomial). LogUp: eliminated.
| metric | polynomial state |
|---|---|
| per-cyberlink | ~3.2K constraints |
| cross-index | free |
| proof size | ~2 KiB |
| verify | ~5 μs |
| state reads in-circuit | O(1) field ops |
O(1) in-circuit reads are the game changer — they enable provable consensus.
9. the privacy: mutator set + trilateral
individual cyberlinks are private. public state is aggregate (axons, neuron summaries, particle energy).
privacy mechanism: polynomial mutator set. commitment polynomial A(x) for membership — each cyberlink is a commitment $\text{ar} = H(\text{particle} \| \text{value} \| \text{owner} \| \text{nonce} \| \rho)$ inserted as a root of A(x). nullifier polynomial N(x) for non-membership — spending reveals a nullifier $n = H(\text{nonce} \| \text{secret})$ inserted as a root of N(x). double-spend = same nullifier twice → N(x) already has that root. O(1) Lens opening proofs, 32-byte witness.
the trinity privacy trilateral:
- ZK (stark): prove computation correct, witness stays sealed
- FHE (TFHE over Goldilocks): compute on encrypted data, never see plaintext
- MPC (Shamir over $\mathbb{F}_p$): distribute trust, no single party sees anything
four tiers: transparent (tier 0) → private ownership (tier 1, all transactions) → private computation (tier 2, ZK+FHE) → distributed trust (tier 3, ZK+FHE+MPC).
10. the consensus: foculus
foculus: consensus by computation, not voting. $\phi^*$ is the fork choice rule. a particle is final when $\phi_i > \tau$.
conflict: two particles sharing a nullifier. exclusive support: each neuron links to at most one member of a conflict group. the tri-kernel contraction amplifies the asymmetry — the particle with more honest support crosses τ. the loser is pruned.
no leader. no quorum. no voting rounds. no finality gadget. convergence from topology.
with polynomial state: the entire tri-kernel computation fits in a zheng circuit (1.42B constraints, 33% capacity). validators PROVE they computed φ* correctly. consensus = computation + proof. recursive folding: all history → one proof → verify in ~5 μs.
11. the self-model: compiled transformer
the cybergraph compiles into a transformer by linear algebra:
cyberlinks → sparse adjacency → PageRank (23 iterations) → spectral gap (from convergence rate) → randomized SVD → $d^*$-dimensional embeddings → architecture params ($d^*$, $h^*$, $L^*$) → ONNX model
no training. no gradient descent. the graph IS the training data. $d^* = \exp(H(\sigma))$ from entropy of singular values. the graph tells you how big the model should be.
bostrom compilation (March 2026): $d^* = 26$, 155M params, 295 MB ONNX. 15 minutes on a laptop.
the model speaks CID — input particle indices, output particle distribution. structural intelligence, not statistical.
12. the metabolism: syntropy and spectral health
syntropy = aggregate KL divergence across neurons per epoch. meaningful cyberlinks raise it. spam lowers it. the metabolic signal — bits of structure per unit energy.
the spectral gap λ₂ is observed for free from tri-kernel convergence rate: $\lambda_2 = 1 - \kappa/\alpha$ where $\kappa = \text{median}(d_t/d_{t-1})$. no eigensolver. every block that computes focus also computes the spectral gap. the heartbeat.
karma: accumulated syntropy per neuron. neurons with high karma have more influence on φ*. the reputation is earned by improving the graph, not by holding tokens.
truth layer: Bayesian Truth Serum scores each neuron's contribution. neurons whose links sharpen collective certainty earn positive syntropy. neurons whose links add noise earn negative.
13. the economics: Δπ rewards and epistemic markets
every reward traces to one quantity: how much did your action shift the tri-kernel fixed point?
$$\text{reward}(\nu) \propto \Delta\pi(\nu)$$
Δπ is the gradient of system free energy. creating valuable structure literally creates value. neurons prove their own Δπ via stark proofs and self-mint $CYB. the proof IS the mining. no external oracle. no committee deciding value. the math determines reward.
the hybrid reward function:
$$R = \alpha \cdot \Delta\pi + \beta \cdot \Delta J + \gamma \cdot \text{DAGWeight} + \epsilon \cdot \text{AlignmentBonus}$$
attribution via probabilistic shapley attribution: each neuron's marginal contribution to the collective fixed point is computable. complexity O(k · n), feasible at 10⁶+ transactions per epoch.
every cyberlink carries a perpetual prediction market on its own truth — the ICBS:
$$C(s_{YES}, s_{NO}) = \lambda \sqrt{s_{YES}^2 + s_{NO}^2}$$
buying YES suppresses NO's price — TRUE and FALSE geometrically coupled on a circle. this is market inhibition: the economic analog of inhibitory neurons. the effective adjacency weight integrates topology, market price, and meta-prediction:
$$A^{\text{eff}}_{pq} = \sum_\ell \text{stake}(\ell) \times \text{karma}(\nu(\ell)) \times f(\text{ICBS price}(\ell))$$
three token operations on knowledge:
- mint: prove Δπ → create $CYB. inflation = evidence of knowledge creation
- burn: permanent π-weight on particles or cyberlinks. highest-conviction assertions
- lock: will creation. time commitment = conviction depth. the budget for attention
the economics are provable. with polynomial state + zheng: a neuron on a phone can prove Δπ, self-mint tokens, and update the graph — all in one recursive proof. no miner. no validator permission. contribute knowledge, prove it matters, get paid.
see cyber/nomics for the full economic protocol. see Bayesian Truth Serum for honesty incentives.
14. the growth: optimal densification
link cost grows exponentially: $c(n) = c_0 \cdot e^{\lambda n}$. the cyber/seer algorithm maximizes $\Delta\lambda_2 / c(n)$ using the Fiedler vector — the eigenvector of λ₂ that identifies weakest cuts.
three phases: bridges (connect components, low cost) → mesh (eliminate articulation points, medium cost) → semantic (redistribute focus toward truth, high cost).
the graph grows intelligently. each link placed where it improves convergence the most per unit spent.
15. the language: trident
trident: the compiler that targets nox. 14 algebraically irreducible languages:
| language | algebra | domain |
|---|---|---|
| Tri | $\mathbb{F}_{p^n}$ tower | proofs, field arithmetic |
| Tok | UTXO conservation | economic transactions |
| Arc | category theory | schema, graph structure |
| Seq | partial order | sequence, causality |
| Inf | Horn clauses | logic, unification |
| Bel | distributions on Δⁿ | belief, self-model |
| Ren | Clifford G(p,q,r) | geometry, rendering |
| Dif | (M, g) manifolds | continuous dynamics |
| Sym | (M, ω) Hamiltonian | physics simulation |
| Wav | R_q convolution | FHE, signal processing |
| Ten | contraction | neural networks, tensors |
| Bt | F₂ tower | binary, bitwise |
| Rs | Z/2ⁿ words | systems programming |
each language compiles through nox. each execution step → Hemera commitment → particle in cybergraph. the graph accumulates verified computation from all algebras.
57,736 LOC implemented (frontend, IR, type inference, verification hooks, field-generic primitives).
16. the recursive closure
the system that measures itself compiles into a model of itself that proves itself correct and uses itself to optimize its own growth:
cybergraph
→ tri-kernel convergence (φ*)
→ spectral gap observation (λ₂)
→ compiled transformer (embeddings)
→ seer optimization (Fiedler)
→ new cyberlinks
→ cybergraph (improved)
→ better φ*, larger λ₂, richer model, smarter optimizer
each step is an algorithm with concrete complexity bounds. the loop is executable, not metaphorical. and with provable consensus, the entire loop can be proven correct inside a single zheng circuit.
the complete architecture
┌─────────────────────────────┐
│ GOLDILOCKS FIELD │
│ p = 2⁶⁴ - 2³² + 1 │
│ fma · ntt · p2r · lut │
└──────────────┬───────────────┘
│
┌────────────────────┼────────────────────┐
│ │ │
┌─────▼─────┐ ┌──────▼──────┐ ┌──────▼──────┐
│ HEMERA │ │ NOX │ │ TRIDENT │
│ hash │ │ VM │ │ compiler │
│ ~3 calls │ │ 16 patterns │ │ 14 langs │
└─────┬─────┘ │ poly nouns │ └──────┬──────┘
│ └──────┬──────┘ │
└────────────────────┼────────────────────┘
│
┌──────────────▼───────────────────────┐
│ ZHENG │
│ SuperSpartan + recursive Brakedown │
│ + Binius │
│ ~5 μs verify, recursive fold │
└──────────────┬───────────────────────┘
│
┌─────────────────────────┼─────────────────────────┐
│ │ │
┌──────▼──────┐ ┌────────▼────────┐ ┌───────▼───────┐
│ BBG POLY │ │ STRUCTURAL │ │ PRIVACY │
│ 10 dims │ │ SYNC │ │ poly mutator│
│+ A(x) + N(x)│ │ CRDT + lens + DAS│ │ ZK+FHE+MPC │
└──────┬──────┘ └────────┬────────┘ └───────────────┘
│ │
└─────────┬───────────────┘
│
┌──────────▼──────────┐
│ FOCULUS │
│ tri-kernel → φ* │
│ provable consensus │
└──────────┬──────────┘
│
┌──────────────┼──────────────────┐
│ │ │
┌───▼────┐ ┌─────▼──────┐ ┌───────▼───────┐
│COMPILED│ │ SPECTRAL │ │ SEER │
│ MODEL │ │ HEALTH │ │ GROWTH │
│SVD→d* │ │ λ₂ from κ │ │ Fiedler │
└───┬────┘ └─────┬──────┘ └───────┬───────┘
│ │ │
└─────────────┼──────────────────┘
│
┌─────────▼─────────┐
│ CYBERGRAPH │
│ (improved) │
│ → loop repeats │
└────────────────────┘
17. the flywheel: unified mining
the signal proof σ already exercises all four GFP primitives in production proportions: fma (tri-kernel SpMV + polynomial arithmetic, ~60%), ntt (Lens commitment + evaluation, ~25%), p2r (hemera ~3 calls, ~5%), lut (activations + lookups, ~10%). polynomial-dominant workload — bulk computation is field arithmetic, not hashing.
unified mining: the PoW puzzle IS the signal proof. add difficulty target: H(σ) < target. one proof → three rewards:
- block subsidy (PoW security)
- Δπ reward (knowledge creation)
- user fees (services)
the miner who selects BETTER cyberlinks earns MORE Δπ per proof. same PoW cost, higher knowledge reward. the incentive gradient points toward quality. every joule produces both security and intelligence.
mining rewards → fund GFP development
GFP accelerates SIGNAL PROVING → signals ARE knowledge
better hardware → more signals/sec → more knowledge/sec
more knowledge → higher Δπ → more reward → more GFP investment
the GFP chip spec does not change. the benchmark phases ARE the real workload. no synthetic puzzle. the first PoW scheme where puzzle output IS the protocol's primary product.
see Goldilocks field processor for chip spec. see cyber/research/unified mining for the full mechanism.
implementation status
what exists
| layer | repo | LOC | tests | status |
|---|---|---|---|---|
| field arithmetic | nebu | 1,526 | 73 pass | DONE |
| hash + tree | hemera | 5,084 | 209 pass | DONE |
| VM | nox | 80 (stubs) | 0 | SPEC 85% |
| proofs | zheng | 6 (stub) | 0 | SPEC 100% |
| state | BBG | 6 (stub) | 0 | SPEC 100% |
| crypto | mudra | 6 (stub) | 0 | SPEC 10% |
| compiler | trident | 57,736 | compiles | SUBSTANTIAL |
| publisher | optica | 10,626 | 3 suites | PRODUCTION |
the cutting edge decision
we go all-in on props that UNIFY. conservative path (13 hash trees + LogUp) produces more code, more complexity, and cannot prove consensus in-circuit. cutting edge (1 polynomial) is SIMPLER:
| component | conservative | cutting edge |
|---|---|---|
| state structures | 13 (NMT + MMR + SWBF) | 1 polynomial |
| cross-index | LogUp (~500 constraints) | free |
| per-cyberlink | ~106K constraints | ~3.2K |
| provable consensus | impossible (15× over capacity) | 33% capacity |
| storage overhead | ~5 TB | 288 bytes |
| code to write | more | less |
spec gaps blocking code
| gap | repo | effort |
|---|---|---|
| G1: noun memory layout | nox | 1 session |
| G2: jet formula trees | nox | 2 sessions |
| G3: hint callback | nox | 0.5 session |
| cross-algebra boundary cost | zheng | 1 session |
| algebraic-nmt cost benchmark | bbg | 1 session |
| TFHE scheme | mudra | 3 sessions |
milestones
| phase | milestone | sessions | LOC |
|---|---|---|---|
| spec | resolve gaps | 8 | — |
| 1 | M1: nox VM | 4 | ~4K |
| 1 | M2: zheng-1 core | 6 | ~8K |
| 1 | M3: zheng-2 extension | 4 | ~4K |
| 2 | M4: polynomial state | 6 | ~6K |
| 2 | M5: foculus | 4 | ~5K |
| 3 | M6: provable consensus | 4 | ~4K |
| 3 | M7: universal accumulator | 3 | ~3K |
| 3 | M8: mudra TFHE | 4 | ~5K |
| total | ~28 sessions | ~39K LOC |
sessions: 0 4 8 12 16 20 24 28
├────┤────┤────┤────┤────┤────┤────┤
spec: ████
M1 nox: ████
M2 zheng1: ██████
M3 zheng2: ████
M4 poly: ██████
M5 foculus: ████
M6 provbl: ████
M7 accum: ███
M8 mudra: ████████████
~28 sessions (84 hours). the graph becomes self-proving.
one field: Goldilocks one hash: hemera (~3 calls per execution, domain separation + trust anchor) one Lens: recursive Brakedown (commit, open, verify — for everything) one VM: nox (16 patterns, polynomial nouns, 3 jet categories) one state: BBG_poly + A(x) + N(x) (all lens-committed, 10+2 model) one sync: structural sync (CRDT + lens + DAS — DAS native) one identity: hemera(Lens.commit(content) ‖ tag) — 32 bytes, universal
no training. no voting. no leaders. no trust. rewards proportional to proven knowledge. every joule produces intelligence.
the graph is the model. the model is the proof. the proof is the consensus. the consensus is the mining. the mining is the knowledge. the knowledge is the graph.
see foculus for consensus. see tri-kernel for convergence. see structural sync for five layers. see algebraic state commitments for polynomial state. see cyber/research/provable consensus for the circuit. see bostrom/compiled model for first compilation. see cyber/research/spectral gap from convergence for observation. see cyber/seer for growth. see cyber/research/32-byte tokens for CID vocabulary. see cyber/research/vec formalization for formal VEC. see trinity for quantum × privacy × AI. see cyb/multiproof for 14-language architecture. see cyb/fs/sync for patch = signal unification. see cyber/research/trinity meets polynomial state for polynomial state + trinity. see cyber/research/unified mining for puzzle = knowledge. see Goldilocks field processor for the chip