Research
Portal AI — papers and technical reports
Every quarter, AI models set new records on benchmarks. In March 2026, Gemini 3.1 Pro scored 94.3% on GPQA Diamond — surpassing human PhD experts by thirty points. The same week, ARC-AGI-3 launched and every frontier model scored below 1%. The benchmarks cannot agree on what the models are. And none of them can tell you whether a single person's life is actually better because of any of it. We introduce VCF (Value Classification Framework) — a framework for measuring AI value in the lived experience of real people. Three orthogonal dimensions: Outcome Primitives (OP), a taxonomy of ten outcome types; Outcome Magnitude (OM), a five-level scale anchored to Human Effort Equivalent; and Evidence Tiers (E), a four-level framework classifying confidence that an outcome occurred. From ~25,000 AI agents, 2,678 users opted in. A 21-day automated classification (March 10–30, 2026) processed 7,718 daily files from 1,305 participants, estimating 17,921 value episodes across all ten outcome types. 41.4% exceeded 8 hours of human effort equivalent. 19.7% left externally verifiable traces in the world — deployed applications, published works, legal filings, profitable trading systems. Cross-validated independently at $9M–$32M in artifact value.
Two people can use the same words and mean different things. Two people can use different words and mean the same thing. The problem of communication has never been the words. Human civilization runs on communication. Every relationship, organization, market, and nation depends on the ability of humans to align their intentions, beliefs, and actions through shared understanding. Yet communication itself — the cognitive operation of constructing, transmitting, and reconstructing meaning between minds — has remained largely outside the scope of formal treatment. This paper proposes a different framing. Communication is the alignment of internal meaning between minds — a lossy, dynamic, measurable process that can be modeled and optimized. Misunderstanding is a system inefficiency, addressable with the same rigor we apply to any other engineering problem. We introduce a formal model of communication as loss minimization between individual meaning spaces — grounded in rate-distortion theory, Bayesian inference, and manifold alignment — situate prior work as partial glimpses of this system, and present first evidence from a live deployment of 1,706 persistent AI agents — analyzed at Day 21 of a deployment that has since grown to over 15,000 — each learning a single human's meaning space through sustained conversational interaction.
Patents
Issued US patents on multi-agent orchestration and verification
A method for identifying and clustering worker agents for processing requests. The core node computes drift metrics, clusters agents by semantic capability, and identifies subsets that are both available and suitable for a given user request.
Context enhancement for multi-agent request processing. Worker agents are clustered by semantic similarity, with drift-aware availability tracking and capability-based routing.
Semantic clustering of worker agents for request processing. Agents are grouped by capability similarity, with drift metrics governing cluster membership and request routing.
User agent verification via embedding similarity and query plan adequacy. A core node evaluates whether a user agent's generated embeddings and query plans meet predefined criteria before approval.
Worker agent verification through capability description testing. The core node generates request-output pairs from the agent's declared capabilities and compares actual outputs against baselines to determine approval.