Aggregation, Proving, and Verification as Irreducible Primitives
Abstract
We argue that all distributed consensus computation decomposes into exactly three irreducible operations: aggregation (combining distributed signals into global state), proving (generating cryptographic evidence of computational correctness), and verification (checking such evidence). We show that what has historically been called “distributed computation” — the paradigm of replicated general-purpose execution pioneered by Ethereum — is a transient fusion of aggregation and verification, necessitated by the absence of practical proof systems at the time of its invention. We trace the evolutionary arc from fused execution through progressive separation to the emerging architecture of accountable aggregation, and demonstrate that this trajectory is driven by an inherent tension between replication and throughput that admits only one resolution. We conclude by characterizing the minimal computational substrate for planetary-scale consensus: specialized aggregation engines united by a universal verification layer, with proving as the bridge between them.
1. Introduction
In 1982, Lamport, Shostak, and Pease established that reliable distributed systems must tolerate Byzantine faults — participants that behave arbitrarily, including maliciously [1]. Their solution required redundant computation: multiple participants execute the same operations independently, and the system accepts the majority result. This foundational insight — that trust requires replication — has shaped every distributed consensus system since.
Four decades later, the landscape of distributed computation appears diverse: Bitcoin verifies transactions, Ethereum executes smart contracts, Solana parallelizes state machines, rollups delegate execution to single parties. These systems are typically classified by their consensus mechanisms (Proof-of-Work, Proof-of-Stake, delegated variants), their execution environments (UTXO, EVM, SVM, Wasm), or their scalability architectures (monolithic, modular, layered).
We propose a different classification — one based not on mechanism but on what kind of work the network actually performs. Under this lens, the apparent diversity collapses into combinations of three operations, and the evolutionary trajectory of the entire field becomes legible as the progressive separation of these operations from their initial fusion.
2. Three Irreducible Operations
2.1 Derivation from First Principles
Consider a set of n participants, each holding private information (balances, preferences, observations, intents), who must agree on some shared state without trusting each other. What operations must the system perform?
First, the participants’ private signals must be combined into a shared outcome. A market price emerges from the intersection of bids and offers. A governance decision emerges from the aggregation of votes. A knowledge graph’s relevance scores emerge from the synthesis of all linking and weighting signals. No single participant can produce this outcome alone — it requires information from many parties. This combination is aggregation.
Second, since participants don’t trust each other (or the aggregator), the correctness of any claimed outcome must be demonstrable. A participant who computes a result must be able to convince others without forcing them to redo the work. This demonstration is proving.
Third, the demonstration must be checkable — efficiently, universally, and without specialized knowledge of the computation being proved. This checking is verification.
These three operations are irreducible in the following sense:
- Without aggregation, there is no shared state — each participant could compute independently and would not need a network at all.
- Without proving, aggregation requires blind trust in whoever performed it, which violates the Byzantine fault tolerance requirement.
- Without verification, proofs are useless — if checking a proof is as expensive as redoing the computation, nothing is gained.
Any distributed consensus system performs some combination of these three. The differences between systems lie in how they implement each operation, who performs each role, and — critically — whether the operations are separated or fused.
2.2 Formal Characterization
Definition 1 (Aggregation). Let $S = {s_1, s_2, \ldots, s_n}ns_ii. An *aggregation function* \A: \mathcal{P}(S) \to \Sigma\sigma \in \Sigma. The function \A has the property that computing \A(S) requires access to all (or a threshold of) signals in \S$.
Definition 2 (Proof). A proof for a computation is an efficiently constructible certificate such that and the time to construct are polynomial in the computation’s complexity, while the time to check is polylogarithmic (or at least substantially sub-linear) in the same.
Definition 3 (Verification). A verification function $V(\pi, y) \to {0, 1}y\pi$. Verification is efficient if its cost is independent of or polylogarithmic in the original computation’s complexity.
The relationship between these operations is asymmetric:
Aggregation requires processing all signals. Proving requires performing the computation plus generating the certificate. Verification requires only checking the certificate. This asymmetry is not incidental — it is the mathematical foundation for the architectural evolution we trace in subsequent sections.
3. The Case Against “Computation” as a Paradigm
3.1 The Ethereum Thesis
In 2013, Vitalik Buterin proposed extending Bitcoin’s limited scripting language into a Turing-complete execution environment — the Ethereum Virtual Machine (EVM) [2]. The thesis was that a “world computer” capable of arbitrary computation would enable a new class of decentralized applications: financial instruments, governance systems, identity management, and coordination mechanisms that Bitcoin’s scripting could not express.
This thesis was operationally correct — the EVM does enable these applications — but taxonomically misleading. It classified Ethereum as a “distributed computer,” implying that its essential function is computation in the general-purpose sense. We argue that this classification obscures what Ethereum actually does.
3.2 What Ethereum Actually Computes
Consider the dominant gas consumers on Ethereum mainnet:
| Application | Gas share | Operation performed |
|---|---|---|
| Decentralized exchanges | ~30% | Combining buy/sell intent into trades and prices |
| Lending protocols | ~15% | Combining deposits and borrows into interest rates |
| Token transfers | ~15% | Checking balances and updating two accounts |
| NFT operations | ~10% | Matching bids, offers, and provenance |
| Bridges | ~8% | Checking claims about external chain events |
| Governance | ~3% | Combining votes into decisions |
| Other DeFi | ~12% | Various financial signal aggregation |
| Miscellaneous | ~7% | Mixed |
Every application that requires a network — that is, every application where value comes from multiple parties interacting — performs aggregation: combining distributed signals (trades, deposits, votes, links) into shared state (prices, rates, decisions, rankings). The applications that don’t require a network (token transfers, access control checks, data integrity verification) perform verification: checking that a locally-produced claim is valid.
No dominant use case performs “arbitrary computation” in the sense that distinguishes a general-purpose computer from a specialized processor. The EVM is Turing-complete, but its actual workload is aggregation and verification — operations that do not require Turing completeness and would be more efficiently served by specialized systems.
3.3 Why the Fusion Occurred
If Ethereum’s actual workload is aggregation + verification rather than general-purpose computation, why was a general-purpose VM built?
The answer is historical, not architectural. In 2015:
-
No practical proof systems existed for arbitrary computation. The first efficient zkSNARK implementations (Pinocchio, 2013 [3]) were academic prototypes, not deployable infrastructure. Without proofs, verification required re-execution — and if you’re re-executing, you need a VM.
-
Aggregation was not recognized as a distinct primitive. The computer science literature discussed consensus, state machine replication, and Byzantine agreement, but did not identify the specific pattern of “combining distributed signals into global state” as a category requiring its own architectural treatment.
-
General-purpose VMs were the available tool. The WebAssembly specification was incomplete. Domain-specific languages for financial aggregation didn’t exist in the blockchain context. The EVM provided a universal substrate because specialized substrates hadn’t been invented yet.
The EVM was the right engineering decision given the technological landscape of 2015. It was the application of an existing paradigm (replicated state machine) to a problem (trustless multi-party coordination) that would later be decomposed into more specific operations.
4. The Replication-Throughput Tension
4.1 The Fundamental Constraint
Byzantine fault tolerance through replication imposes a structural constraint on distributed computation:
Theorem (Replication Bound). In a system where $N nodes each independently execute all operations to achieve BFT consensus, the total computational throughput equals that of a single node, regardless of \N$.
Proof sketch. Each node processes the same sequence of operations to arrive at the same state. Adding nodes increases fault tolerance (the system tolerates more failures) but not throughput (total operations per second remains constant at single-node capacity). The replication factor $N multiplies the total energy expenditure by \N\square$
This is not a conjecture — it is an observable property of every replicated execution system:
| System | Validators | Throughput (TPS) | Throughput per validator |
|---|---|---|---|
| Ethereum L1 | ~900,000 | ~15 | ~15 |
| Solana | ~1,900 | ~4,000 | ~4,000 |
| Cosmos Hub | ~180 | ~10,000 | ~10,000 |
Throughput is determined by single-node capacity, not network size. Solana achieves higher throughput than Ethereum not because it has a better architecture in kind, but because it demands more powerful hardware per node. The throughput curve is set by the most powerful node the network’s economics can sustain.
4.2 The Centralizing Pressure
The replication bound creates a structural pressure toward centralization:
- Users demand higher throughput
- Higher throughput requires more powerful nodes
- More powerful nodes cost more to operate
- Fewer operators can afford the hardware
- The validator set shrinks
- Decentralization decreases
This is not a failure of engineering — it is a mathematical consequence of the replicated execution model. Solana’s trajectory illustrates this: achieving ~4,000 TPS requires validators running on dedicated hardware with 10 Gbps bandwidth, 256 GB RAM, and high-core-count CPUs. The result is ~1,900 validators, predominantly hosted in professional datacenters. Increasing throughput to 40,000 TPS would require 10× more powerful hardware, likely reducing the validator set by an order of magnitude.
The replicated execution model has an inherent ceiling beyond which further throughput gains require sacrificing the decentralization that justified distributed computation in the first place.
4.3 The Resolution
The replication-throughput tension admits exactly one resolution: separate the operations that require replication from those that don’t.
Aggregation must be performed by someone with access to all signals, but doesn’t require $N$-fold replication. A single honest aggregator suffices — if its work can be verified.
Verification can be replicated cheaply. Checking a proof is orders of magnitude less expensive than producing one. $N verifiers checking one proof costs far less than \N$ executors each redoing the entire computation.
Proving is the operation that makes this separation possible. Without proofs, verification requires re-execution, which requires replication, which produces the tension. With proofs, verification requires only proof-checking, which is cheap enough to replicate to arbitrary $N$ without throughput loss.
where $M \ll N\text{Cost}(V) \ll \text{Cost}(A)$. Compare with the fused model:
The separated model scales linearly with verifiers (cheap) while holding aggregation constant. The fused model scales linearly with validators (expensive). For large $N$ — which is precisely the requirement for planetary-scale systems — the separated model dominates.
5. The Role of Redundancy
A natural objection arises: if $N$-fold replication provides fault tolerance, doesn’t reducing replication compromise safety?
The objection confuses execution redundancy with verification redundancy. Both provide Byzantine fault tolerance, but at different costs.
Execution redundancy (the fused model): $Nf < N/3$ Byzantine validators because the honest majority produces the correct result. Every validator must be capable of executing every operation.
Verification redundancy (the separated model): $M provers generate proofs (\M \geq 1, typically small). \Nf < N/3$ Byzantine verifiers (honest majority accepts valid proofs, rejects invalid ones).
The safety guarantees are equivalent — both tolerate up to Byzantine participants in the verification set. The difference is in what each participant must be capable of:
| Property | Execution redundancy | Verification redundancy |
|---|---|---|
| Fault tolerance | Byzantine validators | Byzantine verifiers |
| Per-node cost | Full execution | Proof checking only |
| Hardware requirement | Uniform, high | Asymmetric: provers high, verifiers low |
| Censorship resistance | Any honest validator can include tx | Any honest prover can generate proof |
| Liveness | Requires honest majority of executors | Requires honest prover + honest majority of verifiers |
The key difference is in the liveness requirement: the separated model requires at least one honest prover, whereas the fused model requires an honest majority of executors. This is a weaker liveness assumption (one honest participant vs. $N/2$), but it introduces a different failure mode: if all provers go offline, the system stalls even if all verifiers are honest. In practice, proving is permissionless and economically incentivized, making total prover failure implausible in mature systems.
Redundancy remains valuable and necessary. The question is whether it is achieved by replicating the expensive operation (aggregation/execution) or the cheap one (verification). Proof technology enables the latter.
6. Aggregation: The Irreducible Purpose
6.1 Why Networks Exist
If all computation could be performed by a single party and verified by everyone else, there would be no need for distributed consensus. A bulletin board where provers post proofs and verifiers check them would suffice — no ordering, no consensus, no shared state.
Distributed consensus exists because information is distributed. A market price cannot be computed by any single trader — it emerges from the interaction of all traders’ private valuations. A governance decision cannot be determined by any single voter — it emerges from the aggregation of all votes. A knowledge graph’s relevance scores cannot be assessed by any single participant — they emerge from the synthesis of all contributions.
Aggregation is the irreducible purpose of distributed consensus: the operation that justifies having a network. Remove aggregation, and you remove the reason for the network’s existence.
6.2 Aggregation vs. Computation
Aggregation is a strict subset of computation with specific structural properties:
| Property | General computation | Aggregation |
|---|---|---|
| Input source | Single party | Multiple parties |
| State dependency | Arbitrary | Structured (convergent) |
| Update pattern | Discrete | Often streaming/continuous |
| Output | Arbitrary | Global state reflecting all inputs |
| Requires network? | No | Yes |
General computation includes operations like “compute the 1000th Fibonacci number” — work that a single party can perform and prove without any network interaction. These operations are valuable but do not require distributed consensus. They belong entirely in the proving/verification domain.
Aggregation is specifically the class of computations that require distributed input — where the result is a function of many parties’ private information. This is the class that justifies the complexity and cost of distributed consensus.
6.3 The Universality of Aggregation
Aggregation appears across every domain where distributed systems create value:
Financial aggregation. Order books aggregate bids and offers into prices. Lending pools aggregate deposits and borrows into interest rates. Insurance pools aggregate risk contributions into premiums. In each case, the output (price, rate, premium) is a function of all participants’ private decisions that no single participant could compute alone.
Governance aggregation. Voting systems aggregate preferences into decisions. Delegation systems aggregate trust into representative power. Futarchy aggregates market predictions into policy selections.
Knowledge aggregation. Search engines aggregate linking behavior into relevance scores. Reputation systems aggregate interactions into trust ratings. Citation networks aggregate references into impact metrics. Recommendation systems aggregate user behavior into preference models.
Physical aggregation. Sensor networks aggregate observations into environmental models. Traffic systems aggregate vehicle positions into flow patterns. Energy grids aggregate generation and consumption into pricing and routing.
In every case, the essential operation is the same: many parties contribute signals, and a function combines them into shared state. The function varies by domain (AMM curves, PageRank, voting rules, Kalman filters), but the structure — distributed inputs, global output, continuous updates — is universal.
7. Historical Evolution
The history of distributed consensus computation can be read as the progressive separation of aggregation, proving, and verification from their initial fusion.
7.1 Phase 0: Minimal Aggregation (Bitcoin, 2009)
Bitcoin performs the simplest possible aggregation: ordering. Miners aggregate transactions into blocks, determining which spends occurred first. The aggregation function is trivial — first-come-first-served with fee prioritization — but the consensus about ordering is the core value.
Verification in Bitcoin is efficient by design. Each transaction is independently verifiable: check signatures, confirm inputs exist and are unspent, verify outputs don’t exceed inputs. These are local properties that don’t require knowledge of the full state beyond the UTXO set.
Bitcoin contains no proving in the cryptographic sense (proofs of computational correctness). Proof-of-Work serves a different function — Sybil resistance and leader election — not verification of computation.
Satoshi’s design choice to keep computation minimal was prescient. By limiting the scripting language, Bitcoin avoided the replication-throughput tension entirely. The cost of this choice was limited expressiveness — Bitcoin cannot aggregate complex signals (prices, votes, reputation scores).
7.2 Phase 1: Fused Aggregation (Ethereum, 2015)
Ethereum introduced general-purpose aggregation: any function that can be expressed in EVM bytecode can aggregate signals from any number of parties. This enabled DEXes, lending protocols, governance systems, and the entire DeFi ecosystem.
The fusion was complete — every validator performs every aggregation and every verification by re-executing every transaction. The replication-throughput tension immediately manifested: Ethereum’s throughput (~15 TPS) is determined by single-node execution speed, not network size.
7.3 Phase 1.5: Accelerated Fusion (Solana, 2020)
Solana optimized the fused model through parallel execution (Sealevel), pipelining (Gulf Stream), and hardware acceleration. Throughput improved by ~250× over Ethereum, but the structural model remained identical: every validator re-executes everything.
The acceleration bought time but didn’t resolve the fundamental tension. Solana’s hardware requirements are already substantial, and further throughput gains require proportional hardware escalation.
7.4 Phase 2: Partial Separation (Rollups, 2022)
Optimistic rollups (Optimism, Arbitrum) achieved the first architectural separation: a single sequencer performs aggregation, and the base layer verifies only on dispute. Aggregation is delegated; verification is retained.
The separation is imperfect — the dispute mechanism (fraud proofs) requires the base layer to re-execute the disputed computation, meaning the fused model remains as a fallback. The 7-day challenge window introduces a trust period during which the aggregation is unverified.
ZK rollups (zkSync, StarkNet, Scroll) achieved a cleaner separation: the sequencer aggregates, a prover generates a cryptographic proof, and the base layer verifies the proof without re-executing. No trust window, no fraud game, no fallback to fused execution.
7.5 Phase 3: Intent-Based Aggregation (2024)
Intent-based protocols (UniswapX, CoW Protocol, Anoma) make the aggregation paradigm explicit. Users submit intents (“I want to sell X for at least Y”) rather than execution instructions. Specialized solvers aggregate intents and produce optimal settlements. The network verifies that the settlement satisfies all intents.
This is significant because it drops the pretense of “general-purpose computation” entirely. There is no VM, no bytecode, no replicated execution. There is an aggregation function (intent matching), a prover (the solver), and a verification function (constraint checking). The three operations are explicitly separated.
7.6 Phase 4: Accountable Aggregation (Emerging)
The next frontier applies the proving/verification separation to domains where aggregation is inherently continuous and global: ranking, reputation, attention allocation, reward distribution.
These domains cannot be served by per-transaction proofs (as in ZK rollups) because the aggregation function operates on the entire accumulated state, not on individual transactions. A new cyberlink in a knowledge graph affects the ranking of every connected particle. A new vote in a reputation system affects the trust score of every connected participant. One signal perturbs global state.
The solution is periodic proving: an aggregation engine maintains live state continuously, and at regular intervals produces a cryptographic proof that the state transition since the last checkpoint was correctly computed. Between checkpoints, the aggregator is provisionally trusted. At each checkpoint, trust is replaced by mathematical verification.
This phase is emerging but not yet deployed at scale. It represents the final separation: even the most complex, globally-dependent aggregation becomes cryptographically accountable.
8. Architecture of the Separated Model
8.1 Components
The fully separated model contains three components, each with distinct scaling properties:
SIGNALS ──► AGGREGATION ENGINE ──► PROVING LAYER ──► VERIFICATION LAYER
(few, powerful) (specialized) (universal, light)
Aggregation engine. Maintained by dedicated operators with sufficient hardware to process the full signal stream and maintain global state. Analogous to sequencers in rollups, but for continuous aggregation rather than transaction batching. Few in number (tens to hundreds), with redundancy for liveness and censorship resistance.
Proving layer. Generates cryptographic proofs that the aggregation engine’s claimed state transitions are correct. May be the same operators as the aggregation engine (proving their own work) or independent miners (re-executing and proving independently for rewards). Proof generation is the computational “useful work” — the modern analogue of Proof-of-Work, where energy is converted into trust rather than hash collisions.
Verification layer. Checks proofs. Runs on every participant’s device — phones, sensors, embedded systems, consumer hardware. The cost of verification is polylogarithmic in the computation’s complexity, enabling universal participation without universal computation.
8.2 Role Economics
| Role | Participants | Hardware | Revenue source | Scales by |
|---|---|---|---|---|
| Aggregator | Few (10-100) | High (GPU, high RAM, fast I/O) | Fees + block rewards | Better hardware per node |
| Prover | Many (100-10,000) | High (CPU-intensive) | Proof rewards (mining) | More provers (horizontal) |
| Verifier | Everyone (millions+) | Minimal (any connected device) | Participation rights | More devices (horizontal) |
The critical property: as the network grows, the aggregation cost stays roughly constant (bounded by signal throughput), the proving cost grows linearly with signal volume (but is distributed across provers), and the verification cost per participant stays constant. The system scales horizontally at the verification layer — which is where the participants are.
8.3 Trust Windows
In any practical system, proof generation takes nonzero time. Between the moment the aggregation engine produces a state commitment and the moment the proof is verified, the commitment is provisionally accepted.
Epoch k Epoch k+1 Epoch k+2
├────────────────────────┼─────────────────────────┼──────────
│ Aggregation │ Proof of epoch k │
│ produces σ_{k+1} │ is verified │
│ │ │
│◄── Trust window ──────►│◄── Trustless ───────────►│
The trust window is the cost of separating aggregation from verification in time. Its length is a design parameter: shorter windows require faster proving (higher prover hardware), while longer windows increase the period during which an invalid state could be accepted.
For transaction-level operations (transfers, swaps), ZK rollups achieve near-zero trust windows — proofs are generated and verified within minutes.
For global aggregation operations (ranking, reputation), the trust window is necessarily longer — potentially one epoch (minutes to hours) — because the proof must cover the entire state transition, not a single transaction.
8.4 The Delta Proof Problem
The efficiency of the separated model for aggregation depends on a property we call bounded propagation:
Definition 4 (Bounded Propagation). An aggregation function $A(r, \varepsilon)\Delta s|A(S \cup {\Delta s}) - A(S)|\varepsilonr$ from the affected region.
If the aggregation function has bounded propagation, then the proof of a state transition need only cover the -neighborhood of each new signal, rather than the entire state. The proof size and generation time scale with the number of new signals times the propagation radius, rather than with the total state size.
This property is not universal — some aggregation functions (e.g., naive PageRank with no damping) have unbounded propagation, where a single link change can arbitrarily alter every node’s score. Practical ranking algorithms employ damping factors, convergence bounds, or local update rules specifically to achieve bounded propagation. In the context of the cybergraph, the tri-kernel architecture (diffusion with damping, spring systems with equilibrium, heat kernels with decay) is designed to satisfy this property.
If bounded propagation holds, delta proofs are efficient:
where is the number of new signals, is the propagation radius, and is the graph’s local density. This is dramatically smaller than the full recomputation cost $O(|V| + |E|)$ for large graphs.
9. The Evolutionary Trajectory of Distributed Computation
9.1 The Arc
Viewed through the lens of the three-operation decomposition, the evolution of distributed consensus follows a single trajectory: progressive separation.
| Era | Architecture | Aggregation | Proving | Verification |
|---|---|---|---|---|
| 2009 | Minimal (Bitcoin) | Ordering only | None (PoW is Sybil resistance) | Signature checking |
| 2015 | Fused (Ethereum) | EVM re-execution by all | None | By re-execution |
| 2020 | Accelerated fusion (Solana) | Parallel re-execution by all | None | By re-execution |
| 2022 | Partial separation (Optimistic rollups) | Delegated to sequencer | Fraud proofs (on dispute) | Re-execution (on dispute) |
| 2023 | Clean separation (ZK rollups) | Delegated to sequencer | STARK/SNARK generation | Proof checking |
| 2024 | Explicit separation (Intent protocols) | Solvers aggregate intents | Settlement correctness | Constraint checking |
| 2025+ | Full separation (Accountable aggregation) | Dedicated engines | Periodic STARK checkpoints | Universal proof checking |
The direction is monotonic: from fused to separated. No system has moved in the reverse direction — from separated to fused — because the replication-throughput tension makes fusion increasingly costly as demand grows.
9.2 The Endgame
The trajectory converges toward a stable architecture:
Specialized aggregation engines — each optimized for its domain’s specific aggregation pattern:
- Financial aggregation: order matching, AMM pricing, risk computation (optimized for latency and fairness)
- Knowledge aggregation: ranking, reputation, attention (optimized for convergence and bounded propagation)
- Social aggregation: governance, voting, coordination (optimized for participation and Sybil resistance)
- Physical aggregation: sensor fusion, environmental modeling (optimized for volume and noise tolerance)
Universal verification substrate — a single, domain-agnostic layer that checks proofs from any aggregation engine. Lightweight, universally accessible, horizontally scalable. This is the “trust anchor” of the system — a mathematical judge that confirms or denies the correctness of any claimed aggregation.
Proving as productive work — the economic engine of the system. Provers convert computational energy into cryptographic evidence, earning rewards for making aggregation verifiable. This is the modern analogue of mining: useful work that secures the network, where the “usefulness” is provable computation rather than hash collision search.
9.3 Beyond the Endgame: Ambient Computation
At sufficient scale, the separation becomes invisible. Every device is a potential signal source (contributing to aggregation), a potential prover (generating proofs of local computation), and a verifier (checking proofs from the network). The boundaries between roles dissolve:
- A sensor aggregates its environmental observations, proves the aggregation is consistent with physical models, and submits the proven observation to a larger aggregation network.
- An AI model aggregates training signals, proves its inference follows the claimed architecture, and offers its output as a verified signal.
- A human expresses a preference (a cyberlink, a vote, a bid), which is verified for authenticity and aggregated into the global state.
At this stage, “the blockchain” ceases to be a recognizable artifact. There is no chain, no block, no single network. There is a continuous fabric of aggregation, proving, and verification — three operations performed by billions of devices, composing into a shared model of reality that is trustless not because any entity is trustworthy, but because every claim is provable and every proof is checkable.
10. Implications
10.1 For System Design
The three-operation decomposition suggests design principles:
-
Identify which operation each component performs. If a component aggregates, optimize it for signal throughput and convergence. If it proves, optimize for proof generation speed and size. If it verifies, optimize for minimal resource consumption.
-
Separate operations whenever technology permits. Fusing aggregation and verification (the Ethereum model) should be treated as a technical debt — a temporary expedient that imposes unnecessary costs. As proof systems mature, separation should be pursued aggressively.
-
Design aggregation functions for bounded propagation. The efficiency of the separated model depends on the ability to generate delta proofs. Aggregation functions with unbounded propagation (where one signal change can arbitrarily affect the entire state) resist efficient proving. Functions with bounded propagation enable delta proofs whose cost scales with the change, not the state.
-
Make verification as lightweight as possible. The verification layer is where decentralization lives — where phones, sensors, and embedded devices participate. Every byte of proof size and every millisecond of verification time is a barrier to participation.
10.2 For Evaluation
Existing systems can be evaluated by their position on the separation spectrum:
| System | Separation | Aggregation efficiency | Verification universality | Assessment |
|---|---|---|---|---|
| Bitcoin | High (minimal aggregation) | N/A | High (SPV possible) | Nearly optimal for its scope |
| Ethereum L1 | None (fused) | Low (replicated) | None (requires re-execution) | Transitional |
| Solana | None (fused, fast) | Low (replicated, fast) | None (requires re-execution) | Faster transitional |
| Optimistic rollups | Partial | High (single sequencer) | Partial (on dispute) | Improving |
| ZK rollups | Full (for transactions) | High (single sequencer) | High (proof checking) | Good for transactions |
| Intent protocols | Full | High (solver competition) | High (constraint checking) | Good for discrete matching |
| Accountable aggregation | Full | High (dedicated engine) | High (STARK checkpoint) | Emerging best practice |
10.3 For the Future
The three-operation decomposition predicts:
-
General-purpose distributed VMs will become niche. As specialized aggregation engines and universal proof systems mature, the need for a single “world computer” diminishes. The EVM will persist as legacy infrastructure, but new systems will be designed from the start with separation.
-
Proving will become a commodity market. Proof generation is computationally intensive, parallelizable, and verifiable — the ideal properties for a competitive market. “Proof mining” will replace hash mining as the dominant form of useful computational work in distributed systems.
-
Verification will become ubiquitous. As proof sizes shrink (recursive composition, folding, aggregation) and verification costs drop, every device will be capable of independently verifying the correctness of any computation. Trust will be optional — verification will be default.
-
Aggregation will be recognized as the primary design challenge. With proving and verification increasingly solved by general-purpose cryptographic infrastructure, the differentiating factor between systems will be the quality of their aggregation: how efficiently they combine signals, how robustly they converge, how gracefully they handle adversarial inputs, and how well they scale to planetary input volumes.
11. Conclusion
The distributed computing landscape, viewed through mechanism (PoW vs. PoS, EVM vs. SVM, monolithic vs. modular), appears diverse and fragmented. Viewed through the lens of irreducible operations, it reveals a single evolutionary trajectory: the progressive separation of aggregation, proving, and verification from their initial fusion.
This separation is driven by a mathematical constraint — the replication-throughput tension — that admits no resolution within the fused model. Replicated execution cannot scale to planetary demand without sacrificing decentralization. Separation, enabled by advances in proof systems (STARKs, SNARKs, folding schemes, recursive composition), allows each operation to scale independently: aggregation by algorithm efficiency and dedicated hardware, proving by market competition among provers, and verification by universal lightweight proof checking.
The trajectory converges toward a stable architecture: specialized aggregation engines for each domain, a universal verification substrate accessible to any device, and proving as the productive bridge between them. In this architecture, the network aggregates. It verifies. Proving makes both possible.
Computing is aggregation, proving, and verification. It always was. We are only now building the tools to treat them as such.
References
[1] Lamport, L., Shostak, R., and Pease, M. “The Byzantine Generals Problem.” ACM Transactions on Programming Languages and Systems, 4(3):382–401, 1982.
[2] Buterin, V. “Ethereum: A Next-Generation Smart Contract and Decentralized Application Platform.” Ethereum White Paper, 2013.
[3] Parno, B., Howell, J., Gentry, C., and Raykova, M. “Pinocchio: Nearly Practical Verifiable Computation.” IEEE Symposium on Security and Privacy, 2013.
[4] Ben-Sasson, E., Bentov, I., Horesh, Y., and Riabzev, M. “Scalable, Transparent, and Post-Quantum Secure Computational Integrity.” IACR Cryptology ePrint Archive, 2018.
[5] Nakamoto, S. “Bitcoin: A Peer-to-Peer Electronic Cash System.” 2008.
[6] Bonneau, J., Miller, A., Clark, J., Narayanan, A., Kroll, J.A., and Felten, E.W. “SoK: Research Perspectives and Challenges for Bitcoin and Cryptocurrencies.” IEEE Symposium on Security and Privacy, 2015.
[7] Bünz, B., Bootle, J., Boneh, D., Poelstra, A., Wuille, P., and Maxwell, G. “Bulletproofs: Short Proofs for Confidential Transactions and More.” IEEE Symposium on Security and Privacy, 2018.
[8] Kothapalli, A., Setty, S., and Tzialla, I. “Nova: Recursive Zero-Knowledge Arguments from Folding Schemes.” CRYPTO, 2022.