Skip to content

The Brain: DELTA's End Goal

The Vision

The human brain doesn't process information as flat sequences. It builds, strengthens, and prunes synaptic connections — dynamically constructing the graph over which reasoning flows. Every thought creates new pathways; every pathway shapes the next thought.

DELTA is building toward this: a system that dynamically constructs its own relational graphs and reasons over them, without relying on pre-defined topology or transformer scaffolding.


The Three Paradigms

Paradigm Structure Edges Graph Source Limitation
Transformer Flat sequence Implicit (attention weights) None — reconstructs structure from sequences Quadratic cost to rediscover what a graph already encodes
GNN Static graph Passive scalar wires Pre-given, fixed Can't learn new structure; edges carry no computation
DELTA / The Brain Dynamic graph First-class computational citizens Self-constructed, continuously refined This is what we're building

The key insight: edges should think. In a brain, synapses aren't just wires — they have their own plasticity, their own state, their own role in computation. DELTA's edge-to-edge attention makes edges computational citizens that attend to each other and collectively discover relational patterns that node-only systems miss.

See Visual Explainer for an interactive diagram of this paradigm gap.


What's Been Proven

Six capabilities validate the path toward The Brain:

  • Edge-first attention works — 100% on derived relations via 2-hop edge adjacency (Phase 11), +24% noise robustness over vanilla GNN (Phase 28)
  • Competitive on real KGs — SelfBootstrapHybrid MRR 0.5089, within 0.004 of GraphGPS, beats it on H@10 (Phase 40)
  • Multi-hop advantage accelerates with depth — 5p MRR 0.790 vs GraphGPS 0.690; DELTA is the only model that improves with reasoning depth (Phase 44)
  • Self-bootstrap removes the transformer scaffold — DELTA bootstraps DELTA at 157% of FixedChain (Phase 39)
  • Temperature reveals edge/node asymmetry — Edge attention wants sharper, node wants softer; asymmetric init yields LP MRR 0.4905 (Phases 46–52)
  • Differentiable graph construction works — BrainEncoder (Gumbel-sigmoid edge selection) matches delta_full LP MRR while adding +4.7% H@10 recall from constructed edges (Phases 55–58)
  • Attention dilution is the real scaling bottleneck — At N=5000, 1-layer DELTA surpasses DistMult but 3-layer catastrophically over-smooths. E_adj subsampling is a minor confound (+0.007 MRR); the real issue is attention dilution across 15M+ edge-adjacency pairs (Phases 59–63)

See Key Findings for detailed evidence. See Validation Phases for complete results.


The Capacity Paradox

DELTA-Matched (157K params) beats DELTA-Full (293K params) on multi-hop reasoning. Smaller model, harder task, better result:

Depth DELTA-Matched (157K) DELTA-Full (293K) GraphGPS (228K)
2p 0.758 0.711 0.754
3p 0.753 0.692 0.727
5p 0.790 0.690

The capacity constraint is a feature, not a limitation. The smaller model can't memorize local edge statistics, so it's forced to learn generalizable relational abstractions that compose across hops. This is synaptic pruning — the brain's mechanism of starting with excess connectivity and selectively eliminating connections that don't contribute to function.

DELTA already has the infrastructure for adaptive capacity:

  • PostAttentionPruner — continuous [0,1] importance gates per edge and node, driven by observed attention weights
  • TieredMemory — variational compression (warm tier) and node absorption (cold tier). The graph literally shrinks when cold nodes are absorbed
  • Self-Bootstrap — re-bootstrapping with a smaller architecture that inherits structural knowledge from the larger one
  • LearnedAttentionDropout — per-edge learned dropout probability. Structural edges get low dropout; noisy edges get high dropout

The open question: can DELTA discover its own optimal capacity from data, rather than requiring hyperparameter search?


Roadmap to The Brain

Horizon 1: Core Proven (Phases 41–45) — Complete

Multi-hop compositional advantage validated across depths, seeds, and regularization regimes. Inference cost is deployment-friendly. See Validation Phases.

Horizon 2: Adaptive Architecture (Phases 46–58) — Complete

Learnable temperature revealed the edge/node asymmetry that drives DELTA’s behavior (Phases 46–52). Multi-seed validation (Phase 53) and high-power evaluation (Phase 54) confirmed LP improvements are robust but multi-hop temperature claims are not. BrainEncoder validates differentiable graph construction: MRR 0.4818, H@10 0.8076 — matching delta_full precision while adding +4.7% recall from self-constructed edges (Phases 55–57). Constructor density validated at d=0.01 with 3-seed robustness (Phase 58). Temperature annealing is counterproductive on brain_hybrid.

Horizon 3: Sparse Attention & Full Brain Stack (Phases 59–67+) — Active

KG scaling investigation (Phases 59–63) revealed that attention dilution — not subsampling — is the real bottleneck at scale. At N=5000 nodes, each edge attends to ~15M adjacency pairs; the signal-to-noise ratio collapses. Subsampling ablation (Phase 63) confirmed this: increasing E_adj retention from 23.8% to 100% gains only +0.007 test MRR.

The fix: sparse attention. Rather than attending to all edge neighbors, use top-k or learned sparse patterns so edges attend only to their most informative neighbors. This directly addresses the dilution mechanism.

Concrete phase sequence:

Phase Goal Key Question
64 Top-k sparse edge-to-edge attention Does restricting to k most-relevant E_adj neighbors at N=5000 match or exceed Phase 63 Condition B (MRR ≥ 0.2471) at ≤50% wall time?
65 Full Brain stack activation With sparse attention solving dilution, does enabling PostAttentionPruner + LearnedAttentionDropout + TieredMemory (currently bypassed in Brain mode) improve brain_hybrid MRR?
66 LRA ListOps sequence pilot Can BrainEncoder construct useful relational structure from flat sequential input? This is the critical KG→sequence domain crossing.
67+ Iterative refinement & domain expansion Multi-pass graph construction, temporal reasoning, multi-scale hierarchical graphs

Phase 64 is the linchpin: if sparse attention works, it unblocks both further KG scaling AND the transition to sequence domains (where constructed graphs will also be large).

Horizon 4: Dynamic Reasoning — Future

  • Iterative graph refinement — multi-pass construction where each DELTA pass restructures the graph for the next
  • Temporal reasoning — graphs that evolve over time, with edges that strengthen or decay
  • Multi-scale construction — hierarchical graphs (entity -> concept -> domain) built bottom-up
  • Online learning — graph structure adapts during inference, not just training

Horizon 5: The Brain — Vision

  • Multi-modal graph construction — build relational graphs from text, images, structured data
  • Associative memory — long-term graph state that persists across tasks (like synaptic weights)
  • Compositional generalization — combine known relations to infer novel ones without retraining
  • Autonomous structure discovery — the system discovers what entities and relations exist, not just how they connect

Why Not Just Use Transformers?

Transformers pay a quadratic tax to rediscover structure that a graph already encodes. Every self-attention layer recomputes "who should talk to whom" from scratch. A brain doesn't do this — it has persistent connections that encode which neurons talk to which others.

DELTA's bet is that for relational reasoning — understanding how things connect, not just what things are — operating on explicit relational structure will be fundamentally more efficient than reconstructing it from flat sequences.

The Brain isn't about replacing transformers everywhere. It's about building something better suited for relational reasoning — and scaling it to see how far that advantage extends.


Open Questions

  1. Can sparse attention solve the dilution bottleneck? Phase 63 proved subsampling isn’t the issue; attention dilution across 15M+ pairs is. Top-k or learned sparsity is the proposed fix (Phase 64).
  2. Do the bypassed Brain components (pruner, memory, learned dropout) help when dilution is controlled? These were designed for adaptive capacity but have never been tested with sparse attention (Phase 65).
  3. Can the constructor learn useful graph structure for sequence domains? The SelfBootstrap result (Phase 39) shows DELTA can learn structure from scratch. Can BrainEncoder extend this to non-relational inputs like LRA ListOps? (Phase 66)
  4. Where does explicit structure win vs lose? The hypothesis: relational tasks benefit from explicit graphs, but sequential/generative tasks may not. Where’s the crossover?
  5. Can DELTA scale to millions of entities? Current experiments use ≤15K entities. Sparse attention is a prerequisite.

See Architecture Overview for technical details. See Status & Roadmap for current progress.