Thesis: No single school wins alone. The convergence pattern of 2025–2026 is a hybrid: scaling (connectionist base) + world models (embodied structure) + verifiable reasoning (symbolic scaffold) + alignment (dialectic constitution) + new substrates (organismic hardware). Every frontier lab is now converging toward some weighted combination. The durable disagreements are about weights, not kinds.

  graph TD
      Scale[Connectionist scaling] --> Conv[Convergent 2026 system]
      WM[Embodied world models] --> Conv
      NS[Neuro-symbolic verifiers] --> Conv
      RL[Verifiable-reward RL] --> Conv
      Align[Constitutional alignment] --> Conv
      Interp[Mechanistic interpretability] --> Conv
      Evo[Evolutionary search] --> Conv
      AI2[Active inference hooks] --> Conv
      HW[Alternative substrates] --> Conv
      Conv --> AGI[Candidate AGI systems]
      AGI --> ASI[Path to ASI]
  

8.1 Why no single school wins alone

Pure scaling hits data limits (Era of Experience is a response to exactly this), reasoning ceilings without verifiers, and interpretability walls that become alignment walls. Pure symbolic systems cannot learn from perception. Pure world models, as of 2026, cannot yet match LLMs on text-heavy tasks. Pure active inference is theoretically elegant and empirically behind. Pure evolutionary search is compute-hungry in ways that rival scaling. Pure causal inference is brittle without representation learning.

What works is hybridization. AlphaProof pairs Gemini-based generation with Lean verification and RL. Gemini Deep Think pairs LLM generation with parallel hypothesis search. o3 pairs a pretrained base with deliberative alignment and massive test-time compute. Claude 4.5 pairs scaling with Constitutional AI, mechanistic interpretability, and model welfare. V-JEPA 2 pairs a non-generative latent-space predictor with robot planning. Physical Intelligence π-0.5 pairs a VLM backbone with flow-matching action heads and heterogeneous data. Sakana M2N2 pairs evolutionary merging with quality-diversity archives. VERSES Genius pairs active inference with agentic scaffolding.

The winners in 2026 are all neo-Kantian in the deep sense: they combine empirical scale with architectural priors tailored to the structure of the world.

8.2 Convergence pattern

Five ingredients recur across frontier systems:

  • A pretrained base at frontier scale (LLM, VLM, or world model).
  • Architectural priors suited to the target domain (attention, MoE, SSM, JEPA, liquid, categorical, active-inference modules).
  • Post-training RLVR with verifiable rewards (GRPO or successors).
  • Inference-time deliberation with adjustable compute (o-series, Deep Think).
  • Alignment overlay (Constitutional AI, deliberative alignment, interpretability-informed safety, model welfare).

Divergences are weights, not kinds. OpenAI weights the base heaviest. DeepMind weights deliberation and specialized verifiers. Anthropic weights alignment and interpretability. Meta and World Labs weight architectural priors for the world. Sakana weights evolutionary search. VERSES weights active-inference priors. Extropic and Liquid weight substrate.

8.3 Us, them, and convergence

A recurring motif in 2025–2026 public discourse: biological intelligence (us) vs. silicon intelligence (them) vs. convergence (an intertwined future). Seth's biological naturalism, Friston's free energy principle applied to organisms, and Li's spatial intelligence all argue that intelligence as-we-know-it is bound up with a particular kind of physical being. Sutton's Era of Experience, Kurzweil's Singularity Is Nearer (2024), and the accelerationist fringe argue the opposite: intelligence is substrate-independent and will run away on silicon.

The convergence view — most clearly articulated by Hassabis, Amodei, Bengio, and the Butlin-Long consortium — is that silicon and biological intelligence will interact through a long hybrid period: AI copilots in science, brain-computer interfaces, wet-lab automation (GenBio), robot embodiment (Physical Intelligence), and model welfare programs that treat advanced AIs as potentially morally relevant. The deepest question of 2026 is not which kind of intelligence wins, but what kinds of intelligence can coexist under what institutional arrangements.

8.4 How to read any new AI development

A concise heuristic toolkit. For any new paper, product, or announcement, answer seven questions:

  1. Which school(s) is the authors' philosophical commitment? (Use the Part III map.)
  2. Where in the pipeline does this intervene? (Data, pre-training, architecture, post-training, inference-time, substrate — Part VI.)
  3. What is the architectural prior? (Kantian category: what structure is smuggled in?)
  4. What is the reward signal or objective? (Humean correlation, RLHF, Constitutional, RLVR, free energy, novelty, QD.)
  5. What scaling axis does it claim to open? (Pretraining compute, test-time compute, search compute, evolution compute, energy efficiency.)
  6. What philosophical objection would a principled critic raise? (Scaling → induction problem; world model → evaluation; symbolic → brittleness; active inference → falsifiability; evolutionary → compute cost; interpretability → reductionism.)
  7. Which school does it NOT engage, and is that a principled omission or a blind spot?

Apply this to any 2026 announcement. Marble? Embodied/world-models school, substrate-agnostic, intervenes at pre-training and inference, Kantian prior is 3D/4D spatial structure, objective is reconstruction + consistency, new scaling axis is world complexity, critic raises evaluation difficulty, omits symbolic and active inference. INTELLECT-3? Scaling + RLVR post-training, decentralized compute axis, MoE architectural prior, verifiable reward objective, omits world models and active inference by design.

The map lets you parse announcements without being impressed or dismayed by marketing.

Coda — A personal map for the reader

You are a data science student at Columbia building this into a personal site. The honest subtext of the whole primer: no frontier researcher today is philosophically homeless. Every serious person has a posture, even when they pretend not to. Sutton is a Humean scale maximalist who got quieter about the bitterness as IMO gold arrived. LeCun is a Kantian who walked out of Meta rather than ship another LLM. Hassabis is a scaffolded pluralist who wants to keep every option open. Amodei is a hedging scale-plus-safety pragmatist. Friston is an unrepentant Spinozist. Li is a Merleau-Pontian with a GPU budget. Stanley is a Darwinian. Pearl is a reluctant father figure the field keeps rediscovering.

You get to choose. More usefully: you get to notice you are choosing. That noticing is what this primer is for.

The last heuristic: when you read a new paper and feel either breathless excitement or dismissive contempt, pause and ask which school is provoking the reaction. Then ask which school you have been neglecting. The frontier of AI in 2026 belongs to people who have read deeply enough in more than one tradition to build hybrids that do not yet exist. Philosophy is not the backdrop to this work. It is the design space.