Beyond Scale: The 2025 Landscape of Deutsch-Style AGI Efforts
- exkn
- Jun 15
- 3 min read

Why Deutsch Matters
David Deutsch argues that true progress—scientific, technological, even moral—comes from explanatory knowledge: testable conjectures born out our our minds, that survive relentless criticism. Deutsch-style AGI will be far different than scaled up versions of the current LLM models. It will be sentient and creative in the same ways humans are.
The use cases for the current LLM's are vast and financially enticing so the investment direction of the "AI Boom" is understandable. But I'm interested in what projects are actually working on the Deutsch-style AGI problem, or at least what projects have the potential to converge on it. Here is a chatGPT o3 survey of the current efforts in this direction. It has no idea what it put together here, but in fine LLM style it has synthesized the relevant information into a useful and insightful table.
I'm particularly interested in the Numenta “Thousand Brains” project. I feel like it's on the right track in some important ways.
Facet | What it means |
EK – Explanatory knowledge | Understands reality through causal, criticisable models, not statistical curve-fitting. |
CC – Creative conjecture | Can invent genuinely new ideas, not just remix prior data. |
CF – Critical feedback | Tests and refines conjectures to produce hard-to-vary explanations. |
EC – Error correction | Treats all knowledge as fallible; builds in mechanisms to detect and fix mistakes. |
FW – Free-will-like autonomy | Sets or renegotiates its own goals; can refuse or reprioritise tasks. |
The 2025 AGI Effort Map
Line of attack / Flagship project | Core idea | 2024–25 milestones | Deutsch-facet coverage† | Key pros | Key challenges |
Embodied sensorimotor Numenta “Thousand Brains Project” | Code open-sourced; new non-profit formed | EK: ✔︎ • CC: ✔︎ • CF: ✔︎ • EC: ✔︎ • FW: △ | Biologically grounded; ultra-sparse, low power | Needs rich robotic data; scaling columns unproven | |
Unified curiosity/reward theory; self-rewriting agent jumps SWE-Bench from 20 → 50% | EK: ✔︎ • CC: ✔︎ • CF: ✔︎ • EC: ✔︎ • FW: △ | Formal guarantees; first wild-run self-improvement demos | Proofs explode at scale; reward still hard-wired | ||
Typed knowledge graphs plus neural perception | Hyperon beta due Q4 2025; NARS adds causal abduction | EK: ✔︎ • CC: ✔︎ • CF: ✔︎ • EC: ✔︎ • FW: △ | Human-readable reasoning; modest compute | Perception pipeline, small dev teams | |
Planning & tool-use overlays on giant nets | Web-scale planner prototypes; Olympiad-level maths mode | EK: △ • CC: △ • CF: ✔︎ • EC: △ • FW: ✖︎ | Leverages existing data & infra; ships quickly | Still largely curve-fitting; opaque internals | |
Tasks and agents co-evolve indefinitely | Code-evolving bots double coding benchmark | EK: △ • CC: ✔︎ • CF: △ • EC: ✔︎ • FW: ✖︎ | Generates surprising designs | Fitness ≠ explanations; compute hungry | |
Spike-based hardware & biophysical simulation | 1.15 B-neuron neuromorph opened to researchers | EK: ✔︎ • CC: △ • CF: ✔︎ • EC: △ • FW: ✖︎ | Ultra-low power; test-bed for cortical theory | Programming spikes is hard; leap to reasoning unclear | |
Alignment-by-construction before release | $2 B funding; formal-verification hires | EK: △ • CC: △ • CF: ✔︎ • EC: ✔︎ • FW: ✔︎ | Popperian error-correction at the core; deep pockets | Secrecy slows peer review; proofs still hypothetical |
† Legend — ✔︎ strong evidence △ partial ✖︎ weak/missing
What This Table Shows
Nobody ticks every box—yet.
Thousand Brains and the Gödel-machine line cover the most facets, but both still need a robust preference-formation layer to meet the “free-will” criterion.
LLM-plus-planner projects sprint ahead on benchmarks but remain weakest on explanatory depth and agency.
Convergence signs are real.
System-2 overlays are borrowing symbolic memories and self-critique tools from projects like NARS.
Thousand Brains and Hyperon teams are planning joint robotics benchmarks to anchor sensorimotor grounding.
DGM-style self-improvement is being explored as a plug-in for both hybrid symbolic systems and neuromorphic hardware.
Key open research questions:
Autonomous preference formation – How can an agent develop its own goals instead of maximising a hand-coded reward? (FW)
Public, criticisable explanations – How do we translate internal causal models into statements humans can debate? (EK + CF)
Scalable error-driven learning – Can we keep compute and energy manageable while letting the system learn from every mistake? (EC)
Keeping safety proofs in lock-step – Self-modifying agents evolve; how do formal guarantees keep up? (CC + EC)
Why It Matters
If Deutsch is right, humanity’s biggest breakthroughs will come from ideas that survive criticism, not from larger parameter counts. The projects above are the first serious attempts to code those principles—explanation, conjecture, criticism, correction, autonomy—into silicon. None has cracked the full set, but the diversity of approaches suggests we’re finally exploring the right search space.
Commenti