Piea — The Architecture in Production
Piea is not a concept. It is a running system.
It is an enterprise AI assistant built from scratch on the AIEP evidence substrate. Every response Piea generates is backed by retrieved evidence, cryptographically committed, and independently replayable. Every source is inspected for integrity, every session is persisted in a tamper-evident ledger, and every operation Piea takes is governed by the same constitutional constraints the protocol defines in the abstract.
Piea proves the architecture works.
Piea is live at piea.ai. Enterprise access, API documentation, and vertical specialist modes are available there.
What Piea is
Piea is what an AI assistant looks like when it is built on AIEP rather than on a prediction engine.
A standard AI assistant retrieves text and predicts a confident-sounding answer. It cannot tell you which sources it used. It cannot show you their provenance. It cannot prove the answer has not changed since it was generated. It cannot archive its uncertainty. It cannot replay its reasoning.
Piea does all of these things — not as added features, but as the consequence of running on an AIEP substrate.
| What Piea does | AIEP mechanism |
|---|---|
| Every response cites live sources | Evidence Rail — real-time retrieval with source URLs, hashes, confidence tiers |
| Every answer is cryptographically committed | response_hash — SHA-256 over the answer + evidence set at generation time |
| Low-confidence answers are flagged and archived | Dissent Signal Engine (P126) — governed uncertainty record persisted to ledger |
| Ambiguous questions answered both ways | Semantic Branch Detection (P128) — both valid interpretations answered with the same evidence |
| Full reasoning chain is streamed and stored | Replayable Reasoning Chain (P127) — every step committed, terminal step hash-anchored |
| Untrusted sources are flagged, not silently used | Source Integrity Inspection (P124) — VPN/proxy/no-TLS sources carry warning badges |
| Source provenance governs confidence ceiling | Source Provenance Classification (P124) — five-class taxonomy with ceiling enforcement |
| Documents and URLs are ingested as governed artefacts | Multimodal Ingestion (P119) — PDF, DOCX, plain text → canonical evidence artefact |
| Bulk document sets ingested in single call | Bulk Ingestion (P123) — batch endpoint with delta feed subscription support |
| Evidence sources can be challenged | Evidence Challenge Record (P113) — user-initiated counter-evidence stream |
| Retracted sources cease contributing to confidence | Source Retraction Registry (P114) — propagation to historical sessions |
| Session state persists across channel changes | Multi-Channel Presence Substrate (P116) — Durable Object session continuity |
| Evidence window is qualified before inference | Parametric Unburdening (P117) — five-pass qualification pipeline |
| Prior turns included as governed context | Session Memory Substrate (P118) — KV-backed rolling conversation window |
| Response hash verifiable by third parties | Session Identity (P118) — verify/:hash endpoint with full evidence chain |
| AI agent can take bounded computer-use actions | Computer-Use Execution Surface (P121) — risk-tiered authorisation and action artefacts |
| AODSR tier-1 sources prioritised per domain | Authoritative Open Data Source Registry (P122) — 80+ baseline sources across legislation, treaties, standards |
| Signed audit export on demand | Governed File Output (P120) — Markdown audit pack with full evidence chain |
| Subscription tier governs capability entitlements | Subscription and Billing Protocol (P125) — Stripe-integrated plan lifecycle |
| Answers governed by output self-audit | Meta-Governance (P141) — constitutional self-check before every response is committed |
| Four models cross-checked on every inference | Multi-Model Dispatch — Workers AI (Llama) + GPT-4o + Claude Sonnet + Ollama; 4-dimension consensus weights evidence alignment and model agreement |
| Cross-source synthesis across authority tiers | Synthesis Engine — four modes: consensus, dissent_map, outlier_scan, integration_surface; identifies where Tier-1 sources agree, contradict, or diverge |
| Regulation changes watched and surfaced automatically | Source Monitoring Watchlist — integration_surface mode flags high-value monitoring candidates with provable confidence ceiling |
| Piea operates as domain expert inside other AIEP apps | App Expert Helper (PF-009) — AppContext protocol; Forecast, Hub, and custom apps delegate specialist queries to Piea with live app data |
| Seven vertical specialist modes | Vertical Mode Engine (PF-011) — Construction, Legal, Financial, Planning, Investment, Compliance, Generic; each mode reconfigures sources, system prompt, and confidence routing |
| Fully multi-tenant with role isolation | Multi-Tenancy — TenantStore with five RBAC roles, tenant-scoped evidence ledger, Stripe plan lifecycle |
| Enterprise SSO: Google, Microsoft, Okta, SAML | OIDC/SAML Auth — Cloudflare Access integration; per-tenant SSO provider configuration; white-label UI support |
| Internal databases and feeds as governed evidence | Data Connectors — PostgreSQL, MySQL, REST API, SharePoint, S3/R2, CSV/JSON; internal results committed to Evidence Ledger identically to web artefacts |
| Push responses to Slack, Teams, or any webhook | Integrations — IntegrationRegistry; events: piea.response, piea.source.drift, piea.export.ready, piea.session.resumed; signed outbound payloads |
| TypeScript SDK for external developers | @piea/integrations — PieaClient with evidence rail access built in |
| Semantic source memory across sessions | Vectorize — piea-sources index; cosine-similarity source routing at query time |
| Autonomous discovery of relevant sources | Source Discovery Mode — Piea searches for and proposes new sources; admin approves before they enter the retrieval pipeline |
AIEP is not the model — it is the evidence layer around the model
Piea uses a large language model to generate text. That is not AIEP’s claim.
Every capable AI assistant uses a language model. The model produces text. The text can be good or bad, accurate or hallucinated, useful or misleading. Model quality is a real dimension — but it is not what AIEP governs.
AIEP governs what wraps the generation. Before the model sees a question, Piea has already:
- retrieved live sources and committed each one to an evidence chain (P37)
- inspected every source’s network path for integrity flags (P124)
- qualified the evidence window to remove noise (P117)
After the model generates a response, Piea has:
- bound the response to its evidence set with a cryptographic commitment (R8)
- checked whether the evidence was sufficient — and if not, emitted a dissent record (P126)
- stored the full reasoning chain in a form any third party can replay (P127)
The model is the generation engine. AIEP is the governed evidence substrate the model operates on.
When evaluating Piea against other AI assistants, the relevant question is not “which model generates better text?” — it is “which system can prove what sources it used, verify they were not tampered with, and produce a hash any third party can independently check?” No other system answers that question. Piea does — as a structural property of running on an AIEP substrate, not as a claimed feature.
What Piea implements
Piea is a direct production implementation of the AIEP Piea Surface patent cluster (P116–P128). It is the reference implementation of what governed AI inference looks like at the application layer.
Cryptographic spine
Every Piea response is built on GENOME R1–R8 — the eight canonical primitives:
- R1
canonical_json— deterministic serialisation of every artefact - R2–R3
sha256— hash binding at every step - R4
concat_hash— chain construction across evidence items - R5
evidence_commitment— session evidence set committed before answer generation - R6
lifecycle_hash— lifecycle event binding - R7
negative_proof_hash— absence proven, not merely asserted - R8
response_commitment— the answer itself, hash-bound to its evidence
No response can exist in Piea’s ledger without a valid R8 commitment over its evidence chain.
The Evidence Rail
The Evidence Rail is Piea’s live audit surface — shown as a persistent right panel in the UI. For every response, it displays:
- Source URLs with metadata
ContentHash(SHA-256) where available — detects alteration since retrieval- Confidence tier:
verified·qualified·unverified·community - Source integrity flags: VPN · Relay · No TLS · Geo-Restricted · Stale
- Time since retrieval (
retrieved_at)
The evidence rail is not a footnote. It is the primary governance surface. Every source that influenced a response is visible, inspectable, and linkable.
Dissent Signal Engine (P126)
When Piea’s retrieved evidence is insufficient to meet the confidence threshold, it does not silently answer with reduced confidence. It generates a DissentSignal — a governed record containing:
- The triggering question
- The number of sources retrieved
- The confidence threshold that was not met
- The specific reason for dissent
The dissent record is persisted to the evidence ledger. It is not a warning message. It is a first-class artefact in the session’s evidence history.
Replayable Reasoning Chain (P127)
Piea streams its reasoning steps to the UI before the final answer — over the same SSE channel. Each step is a structured ReasoningStep:
- Query classification
- Source identification
- Evidence retrieval and qualification
- Cross-source synthesis
- Confidence determination and response commitment
The terminal step (step 5) carries the response_hash as its completion signal. The full reasoning chain is stored as a first-class field of the session record — independently replayable from the stored evidence set.
Semantic Branch Detection (P128)
When Piea detects a semantically ambiguous query — one that is valid under two distinct interpretive frameworks — it does not ask for clarification. It answers both interpretations simultaneously, each grounded in the same retrieved evidence set.
Each SemanticBranch carries:
- The interpretive framework it applies
- A complete answer grounded in the shared evidence
- A confidence tier for this specific interpretation
- An
is_primaryflag preserving the conventional reading
Both answers are persisted in the session record. This is not AI equivocation. It is a governed protocol response to genuine interpretive ambiguity — each branch independently auditable.
The infrastructure
Piea runs entirely on Cloudflare’s edge:
| Component | Platform | Role |
|---|---|---|
| Chat UI | Cloudflare Pages (React 18) | Evidence Rail, Dissent, Reasoning Replay, Semantic Branches, Compare, Groups |
| API | Cloudflare Workers (Hono.js) | Evidence retrieval, session governance, ingestion, synthesis engine |
| Core agent | @piea/agent package | Evidence Rail orchestration, R1–R8, all spec implementations |
| Evidence Ledger | Cloudflare D1 (dual-ledger) | Persistent session + evidence store + immutable audit ledger (P80) |
| Session store | Cloudflare KV | Evidence artefact cache (P118) |
| Presence substrate | Cloudflare Durable Objects | Substrate continuity (P116) |
| Primary LLM | Cloudflare Workers AI | Llama-3.3-70B-fp8, routed via P117 evidence window |
| LRM synthesis | OpenAI GPT-4o / Claude Sonnet / Ollama | Multi-model 4-dimension consensus for analytical queries |
| Semantic memory | Cloudflare Vectorize | piea-sources index — BGE-base-en-v1.5 embeddings |
| Mirror artefacts | Cloudflare R2 | Governed file output (P120), AIEP Mirror pages |
| Integrations | Slack / Teams / Webhooks | @piea/integrations — signed outbound event dispatch |
| Multi-tenancy | @piea/tenancy | Tenant isolation, RBAC, OIDC/SAML SSO |
| Data connectors | @piea/data-connectors | PostgreSQL, MySQL, SharePoint, REST API, S3/R2, CSV → Evidence Ledger |
Every component is edge-deployed. No centralised server. No single point of failure. The substrate continuity layer (P116) maintains reasoning state across channel changes without data loss.
Multi-model dispatch
Piea does not route every question to a single model. Every query is classified into one of eight types (factual, procedural, analytical, generative, administrative, domain, conversational, smalltalk) and dispatched accordingly:
- Factual / procedural → Workers AI (Llama-3.3-70B) against retrieved evidence
- Analytical / complex → Multi-model LRM: Workers AI + OpenAI GPT-4o + Claude Sonnet + Ollama run in parallel; a four-dimension consensus weights evidence alignment, source completeness, claim overlap, and model agreement
- Domain / internal → custom sources only; no substrate legislation injected
- Conversational → no sourcing; direct response
The LRM consensus produces a single governed answer synthesised from all model perspectives — not the output of one model but a hash-committed synthesis of their agreement. Dissent zones where models diverge are surfaced to the evidence rail.
Cross-source synthesis engine
Beyond answering individual questions, Piea can run a synthesis operation across the full evidence source graph for any topic. Four modes:
| Mode | What it produces |
|---|---|
consensus | Claims that Tier-1 and Tier-2 sources agree on, with confidence and supporting tiers |
dissent_map | Points of active contention between authoritative sources, with severity rating and confidence impact |
outlier_scan | Lower-tier signals that diverge from the Tier-1/2 consensus — potential leading indicators |
integration_surface | Specific questions Piea can answer with provable high confidence for a given organisation or domain, with monitoring candidates flagged |
The integration surface scan is particularly useful for compliance and investment teams: it identifies exactly where Piea can provide evidence-grounded answers without approximation — and which regulatory changes are worth adding to a monitoring watchlist.
Vertical specialist modes
Seven domain modes reconfigure the full evidence pipeline — sources, system prompt, and confidence routing — without separate applications:
| Mode | Primary sources | Domain |
|---|---|---|
| Construction | HSE, Legislation.gov.uk, RICS, Planning Portal | CDM, NEC, JCT, building regulations |
| Legal | BAILII, Legislation.gov.uk, Supreme Court, EUR-Lex | Primary law, case law, statutory instruments |
| Financial | FCA, Bank of England, HMRC, Companies House, SEC | FRS 102, IFRS, financial regulation |
| Planning | Planning Portal, NPPF, GOV.UK | Local plans, EIA, development management |
| Investment | FCA, BoE, HMRC, ONS, BIS, SEC/EDGAR | Monetary policy, market data, gilt yields |
| Compliance | ICO, FCA, HSE, Legislation.gov.uk | UK GDPR, AML, sector regulation, enforcement |
| Generic | Full 80+ source baseline | Any topic |
Enterprise: multi-tenancy, SSO, and data connectors
Piea is a production enterprise system. It runs with full tenant isolation from day one:
Multi-tenancy and RBAC — each tenant has a scoped evidence ledger, five RBAC roles (super_admin → viewer), and an independent subscription lifecycle. White-label UI is available at enterprise tier.
SSO — OIDC and SAML supported: Google, Microsoft (Entra ID), Okta, and Cloudflare Access JWT. Per-tenant SSO provider configuration. Credentials stored in Cloudflare Secrets — never in D1.
Internal data connectors — PostgreSQL, MySQL, REST API, SharePoint, S3/R2, CSV/JSON feeds. Internal query results are committed to the Evidence Ledger with the same artefact schema as web sources. Piea treats internal and external evidence identically — the retrieval pipeline cannot distinguish origin.
Integrations — Slack, Microsoft Teams, and generic webhooks with signed payloads. Five event types: piea.response, piea.source.added, piea.source.drift, piea.export.ready, piea.session.resumed. TypeScript SDK (@piea/integrations) for external developers.
App Expert Helper protocol (PF-009)
Any AIEP application can call Piea as a domain expert helper using the AppContext protocol. The calling app prepares:
live_context— pre-fetched app data (project summary, snag list, positions, etc.)mode— vertical specialist mode to activatesource_priority— domain-relevant URLs to boostexpertise_kb— static KB chunks committed to the evidence rail
Piea grounds its answer in the app’s live data along with retrieved external evidence — producing an answer that knows both the current state of the application and the current state of the regulatory environment. AIEP Forecast uses this to give Piea live project and CRM context. Any AIEP-compliant application can do the same with no schema changes.
The critical claim is not that Piea is a good AI assistant. The critical claim is that an AI assistant built on AIEP cannot behave like a hallucinating prediction engine — not because it is instructed not to, but because the architecture prevents it.
A prediction-based AI can always generate a confident answer with no evidence. Piea cannot. Every response requires an evidence commitment before generation. An empty evidence set produces a dissent record, not a fabricated answer.
A prediction-based AI cannot prove its answer has not changed. Piea can. The response_hash is a cryptographic commitment over the answer and its evidence — tamper-evident and independently verifiable.
A prediction-based AI cannot replay its reasoning. Piea can. The full reasoning chain is stored as structured artefacts in the same session store as the response. Any session can be audited — step by step, source by source — from the ledger record.
This is what makes Piea evidence for AIEP, not just an application of AIEP. Piea demonstrates that the protocol is complete enough to build a real production system on — and that building on it produces properties no prediction-based architecture can match.
Try Piea
Piea is live at piea.ai. Enterprise access, API documentation, and vertical specialist modes are available there.
Related
- Architecture — the seven-layer AIEP stack Piea implements
- Evidence Layer — normalisation, stitching, and recall
- GENOME SDK — the cryptographic kernel Piea runs on
- Showcase — the full repository and test coverage record
- Patents — P116–P128: the Piea surface patent cluster