Research

Portal AI — papers and technical reports

Hamudi Naanaa, Vitaliy Soultan, Volodymyr Panchenko
March 2026 · Portal AI · Everwhy AI
Every quarter, AI models set new records on benchmarks. In March 2026, Gemini 3.1 Pro scored 94.3% on GPQA Diamond — surpassing human PhD experts by thirty points. The same week, ARC-AGI-3 launched and every frontier model scored below 1%. The benchmarks cannot agree on what the models are. And none of them can tell you whether a single person's life is actually better because of any of it. We introduce VCF (Value Classification Framework) — a framework for measuring AI value in the lived experience of real people. Three orthogonal dimensions: Outcome Primitives (OP), a taxonomy of ten outcome types; Outcome Magnitude (OM), a five-level scale anchored to Human Effort Equivalent; and Evidence Tiers (E), a four-level framework classifying confidence that an outcome occurred. From ~25,000 AI agents, 2,678 users opted in. A 21-day automated classification (March 10–30, 2026) processed 7,718 daily files from 1,305 participants, estimating 17,921 value episodes across all ten outcome types. 41.4% exceeded 8 hours of human effort equivalent. 19.7% left externally verifiable traces in the world — deployed applications, published works, legal filings, profitable trading systems. Cross-validated independently at $9M–$32M in artifact value.
Hamudi Naanaa, Volodymyr Panchenko
March 2026 · Portal AI
Two people can use the same words and mean different things. Two people can use different words and mean the same thing. The problem of communication has never been the words. Human civilization runs on communication. Every relationship, organization, market, and nation depends on the ability of humans to align their intentions, beliefs, and actions through shared understanding. Yet communication itself — the cognitive operation of constructing, transmitting, and reconstructing meaning between minds — has remained largely outside the scope of formal treatment. This paper proposes a different framing. Communication is the alignment of internal meaning between minds — a lossy, dynamic, measurable process that can be modeled and optimized. Misunderstanding is a system inefficiency, addressable with the same rigor we apply to any other engineering problem. We introduce a formal model of communication as loss minimization between individual meaning spaces — grounded in rate-distortion theory, Bayesian inference, and manifold alignment — situate prior work as partial glimpses of this system, and present first evidence from a live deployment of 1,706 persistent AI agents — analyzed at Day 21 of a deployment that has since grown to over 15,000 — each learning a single human's meaning space through sustained conversational interaction.

Patents

Issued US patents on multi-agent orchestration and verification

US12407510B2 · Issued Sep 2, 2025
A method for identifying and clustering worker agents for processing requests. The core node computes drift metrics, clusters agents by semantic capability, and identifies subsets that are both available and suitable for a given user request.
US12395337B2 · Issued Aug 19, 2025
Context enhancement for multi-agent request processing. Worker agents are clustered by semantic similarity, with drift-aware availability tracking and capability-based routing.
US12265856B1 · Issued Apr 1, 2025
Semantic clustering of worker agents for request processing. Agents are grouped by capability similarity, with drift metrics governing cluster membership and request routing.
US12260005B1 · Issued Mar 25, 2025
User agent verification via embedding similarity and query plan adequacy. A core node evaluates whether a user agent's generated embeddings and query plans meet predefined criteria before approval.
US20250088357A1 · Published Feb 25, 2025
Worker agent verification through capability description testing. The core node generates request-output pairs from the agent's declared capabilities and compares actual outputs against baselines to determine approval.