semantic language for neurons over the cybergraph. whitepaper: neural language for superintelligence

convergent successor for both formal and natural languages

meaning is defined by cyberlinks — structure emerges from how agents link particles

part of the soft3 stack, running on Bostrom alongside the tru

the language of egregore: meaning emerges from how many neurons independently structure knowledge

why a new language

  • formal languages (type theory, programming languages) achieve precision through rigid syntax but cannot scale to 10¹⁵ particlesGoedel proved no sufficiently powerful formal system can be both complete and consistent (the Goedel prison)
  • natural languages solve expressiveness through ambiguity but are computationally intractable for precise reasoning
  • neural language collapses the distinction between language and knowledge: meaning is an eigenvector of the attention graph
property formal natural neural
precision absolute approximate emergent
expressiveness limited by grammar unlimited by ambiguity unlimited by topology
ambiguity impossible context-dependent structural via tri-kernel
authority central designer speech community collective neurons
evolution versioned drift continuous via focus dynamics
machine readable yes partially via NLP natively
human readable requires training natively via cyb interface
verification proof systems social consensus stark proofs
substrate strings sound/text cybergraph

patterns

  • semcon

    • semantic conventions — mutual agreements to use the same particles for structuring thought
    • the grammar of the graph
    • a semcon is a smart contract that creates cyberlinks according to convention — invocation produces well-formed graph structure
    • the neuron provides intent, the semcon handles structural correctness
    • bootloader semcons installed at genesis: TRUE, FALSE — the epistemic coordinates from which all meaning derives
    • emergent semcons discovered by the network: is-a, follows, causes, contradicts, part-of, see-also
    • semcon hierarchy emerges from topology: structural → domain-specific, epistemic → modal, temporal → causal, social → evaluative
    • the tri-kernel reveals semcons: diffusion identifies high-betweenness bridges, springs reveal stable structural positions, heat modulates attention by adoption weight
  • sentence

    • ordered instruction set of cyberlinks — a batch packed into a single transaction
    • the transaction boundary defines the utterance. order within the batch encodes grammar
    • transaction-atomic semantics: every transaction is a linguistic act
    • sentence types by topological signature: assertion (chain → TRUE), query (open-ended chain), instruction (temporal sequence), argument (branching to TRUE/FALSE), definition (star pattern), narrative (temporally ordered chain)
    • sentences compose through shared particles — creating linkchains the tri-kernel can discover
  • motif

    • geometric expression of meaning — recurring subgraph patterns that encode relationships beyond single cyberlinks
    • the morphemes of neural language
    • triadic closure: if A links B and B links C, A linking C completes a trust/relevance triangle
    • co-citation: multiple neurons linking the same pair signals consensus
    • star: one particle linked by many signals centrality or definitional importance
    • chain: sequential links encoding transitive, causal, or narrative relationships
    • diamond: convergent-divergent pattern — multiple paths between endpoints signals robust relationship
    • motif algebra: concatenation (transitive reasoning), nesting (hierarchical abstraction), intersection (cross-domain bridges), complement (knowledge gaps)
  • name

    • deterministic resolution of a cyberlink: given from, return exactly one to — the latest particle linked by the owning neuron
    • standard resolution is probabilistic (ranked candidates by cyberank); the ~ prefix signals deterministic resolution
    • ~neuron/path turns the cybergraph into a dynamic file system — every neuron maintains a namespace rooted at ~
    • the same mechanism underlies file systems, DNS, ENS — all are dynamic pointers where a fixed label resolves to a mutable target
    • a semcon: structural convention distinguishing addressing from search
  • cyberlink as particle

    • a link stored as a particle itself, enabling links about links — meta-knowledge
    • the recursion that makes the language expressively complete
    • enables: negation, qualification, provenance, annotation
    • the language can talk about itself

semantic core

  • the dynamic vocabulary of the network — top particles by cyberank
  • defined by focus distribution: SemanticCore(k) = top k particles by π
  • current core shaped by bostrom bootloader
  • explore at cyb.ai/particles
  • properties: dynamic (evolves with attention), convergent (tri-kernel guarantees stability), stake-weighted (resistant to spam), verifiable (stark proofs)
  • dynamics mirror natural language: neologism (new concepts enter), semantic drift (meaning shifts through topology change), semantic death (focus drops below threshold), semantic birth (bursts of link creation)

linkchains

  • sequences of cyberlinks that form paths of meaning through the cybergraph
  • a → b → c encodes transitive relationship: if a relates to b and b relates to c, the chain implies a relates to c
  • the tri-kernel discovers these implicit paths through diffusion
  • the springs kernel enforces structural consistency across chains — contradictions create tension resolved by dampening
  • properties: length (shorter = stronger), width (parallel paths = robust), weight (product of edge weights)
  • linkchains are the inference mechanism: sentences are explicit statements, linkchains are implicit conclusions

relationship to the stack

formal properties

  • ambiguity resolution: topology around a particle disambiguates meaning computationally — springs detect polysemy as high tension, heat concentrates on contextually appropriate meaning
  • compositionality: meaning of complex expression derivable from parts and their structural arrangement — computed by tri-kernel without explicit composition rules
  • convergence: inherits from the Collective Focus Theorem — unique stationary distribution π* guarantees the network's collective understanding converges
  • expressiveness: semantically complete — can encode propositional logic, predicate logic, modal logic, temporal logic, fuzzy logic, and natural language semantics. can also express collective confidence distributions, continuous semantic distance, and knowledge topology metadata

evolution phases

implementation

connections to linguistic theory

Dimensions

trident/src/ir/tir/neural
neural
trident/reference/neural
Neural Optimizer GNN encoder + Transformer decoder (~13M parameters) that compiles TIR to TASM. Operates at the TIR→TASM boundary — the only non-deterministic stage in the pipeline. Architecture **Encoder:** 4-layer GATv2 (Graph Attention Network v2), d=256, d_edge=32. Encodes a `TirGraph` (typed…
trident/src/neural
neural

Local Graph