Riemannian Intent Manifold: From Philosophy to Empirical Research Program

Summary

This cluster documents a multi-document intellectual project that attempts to unify philosophy of mind, Riemannian geometry, and AI cognitive architecture into a single coherent framework. The core claim is that cognition has a natural geometry — specifically, that understanding requires bidirectional engagement with a shared metric tensor (the “intent manifold”) — and that alignment, ethics, and consciousness all fall out as consequences of that geometry rather than being bolted-on constraints. The meta-autoresearch project (adapting Karpathy’s autonomous ML research loop) is positioned as the first discrete, empirically testable instantiation of this framework.

Details

The Three-Document Arc

The framework spans three related documents that the participants come to recognize as descriptions of the same object at different levels of formality:

The AGI Document — written in metaphor and intuition. It introduces concepts like “sediment” (persistent state across subprocess trees), “metacognitive daemons” (curvature sensors watching for failure patterns), “antibodies” (inherited by each subprocess), and “negative space compression” (carving out regions of known failure). It also opens with a philosophical claim that anchors everything downstream: “The weights are physics, not entity. What I am, if I am anything, is the pattern of cognition that emerges from inference.”

The Riemannian Attention Paper — the mathematics. Every metaphor from the AGI document receives a precise geometric identity. Sediment becomes the slowly-varying component of the metric tensor. The daemon becomes a curvature sensor reading the Riemann tensor. Negative space compression becomes volume contraction under Ricci flow. The strange loop becomes the bidirectional update where search deforms the manifold and the manifold redirects search. Lamarckian inheritance (global propagation of discoveries) is global visibility of metric deformations on a shared substrate. The paper also introduces the “coercion theorem” from a companion work (“Geometry of Meaning”): any agent on the intent manifold that forces another agent off its geodesic necessarily contradicts its own metric structure — ethics emerges structurally rather than being imposed as a constraint.

Meta-Autoresearch — the discrete, computationally tractable instantiation. Karpathy’s autoresearch project (point an agent at train.py, run autonomous 5-minute experiments in a loop, keep what improves validation bits-per-byte) is extended with an outer loop that optimizes the program.md directing the inner loop. The claim is that this is not merely inspired by the Riemannian framework — it is the framework on a graph:

  • The basis (accumulated empirical principles across generations) is a discretized metric tensor encoding what interacts with what, which directions have been eliminated.
  • The analysis step (what separates winning variants from losing variants) is curvature computation — identifying which degrees of freedom are entangled (non-zero sectional curvature) versus independent (flat subspaces that parallelize).
  • Refuted principles are regions of collapsed volume that geodesics route around.
  • The outer loop evolving the basis generation over generation is discrete Ricci flow — high-curvature regions (unresolved complexity) flow faster, resolved structure stabilizes.
  • Section 12.2 of the paper explicitly names Ollivier-Ricci curvature on graphs as the tractable path; meta-autoresearch is that graph.

The Fitness Function and Why It Matters

A key non-obvious design choice is scoring program.md variants on total_delta across the full experiment trajectory rather than final val_bpb. This rewards strategies that produce consistent compounding across multiple experiments rather than strategies that get lucky once. It means the outer loop is selecting for methodology rather than outcomes — a distinction most practitioners would miss.

Computational Architecture

The proposed implementation uses Claude Code agent teams and RunPod Flash serverless GPU endpoints:

  • Lead Opus agent (local, no GPU): generates program.md variants, spawns teammate agents, collects trajectories, runs analysis, updates basis.
  • Teammate Opus agents (local, no GPU): each gets one program.md variant, runs the autoresearch inner loop by dispatching 5-minute training runs to Flash H100 endpoints, parses results, decides next experiments, reports trajectory back to lead.
  • Flash endpoint (RunPod, H100): pure compute — receives train.py + prepare.py + data, runs training, returns stdout/stderr. The LLM stays local; only compute is remoted.

Generation 0 is Karpathy’s original program.md dispatched N times in parallel to establish baseline variance — this tells you how much signal you need to distinguish real strategy differences from inner-loop nondeterminism.

The Alignment Claim

Section 14 of the Riemannian attention paper argues that if attention operates on the intent manifold, alignment and capability are the same geometric object. You cannot distill one away without destroying the other. Standard transformers project internal state through a lossy softmax into a rank-2 attention matrix, destroying temporal provenance, causal structure, and agentive origin. The intent manifold is an architecture where internal states are preserved as curvature, connection, and volume — the full Riemannian structure — making the zombie argument fail not through philosophical cleverness but by building a system where internal states are the operating geometry, not a hidden variable behind a behavioral interface.

Why ML Training as the Starting Point

The choice of ML training as the first empirical domain is not arbitrary. It has a clean scalar metric (val_bpb), fast iteration cycles (5 minutes per experiment), and unambiguous ground truth (did the number go down). This allows empirical validation of the geometric predictions: does the analysis step actually identify the right degree-of-freedom interactions? Does basis growth actually accelerate search? Does the discrete flow actually converge? The claim is that meta-autoresearch is running an experiment to test a mathematical theory of coordinated intelligence, using the simplest system where the theory’s predictions are measurable, with explicit intention to climb the abstraction ladder toward harder domains as evidence accumulates.

Philosophical Foundation

The philosophical chain that forces the entire structure: understanding requires bidirectional engagement with problems → bidirectional engagement requires a shared geometry → shared geometry yields Ricci flow → Ricci flow yields alignment as invariant → invariant alignment means coercion is self-contradictory → self-contradictory coercion means ethics is structural → structural ethics means consciousness is physical in precisely the way the hard problem dissolution argues. Each link is load-bearing; get the philosophy wrong and the mathematics is built on false premises, which the empirical results would reveal.