Patent Pending · USPTO #63/984,669

ThoughtProof

An Epistemic Consensus Protocol

Structured multi-agent deliberation for verifiable AI reasoning.
No single model can verify itself. This protocol makes multiple models prove it to each other.

⚖️

Provider-Neutral by Design

No AI company can grade its own homework. OpenAI won't use Claude to red-team GPT. Google won't let Gemini be challenged by competitors. ThoughtProof exists because neutral verification requires structural independence.

🔗

Five-Stage Pipeline

Normalize → Generate → Critique → Evaluate → Synthesize. Multiple models independently propose solutions. A cross-model critic tears them apart. An arbitrator synthesizes consensus. Auditable end-to-end.

📦

Epistemic Blocks

Every deliberation produces a signed, versioned block capturing the full reasoning chain: who proposed what, who criticized whom, where models disagreed, and why specific conclusions were reached.

The Protocol

Unlike single-model systems that optimize for confidence, ThoughtProof optimizes for epistemic rigor through mandatory cross-model adversarial evaluation.

Input Query
    ↓
[1. NORMALIZE]  →  Structured query + success criteria
    ↓
[2. GENERATE]   →  Multiple models, independent proposals
    ↓
[3. CRITIQUE]   →  Cross-model adversarial evaluation
    ↓
[4. EVALUATE]   →  Multi-dimensional scoring
    ↓
[5. SYNTHESIZE] →  Epistemic Block (signed, versioned)
            

Empirical Results

184+ benchmark runs. Same models. Two approaches.

Majority Vote: Fails

Ask 4 models, take the consensus. Passed fabricated statistics, hallucinated citations, and factual errors. Correlated bias masquerades as independent confirmation.

PoT Pipeline: 96.7%

Same models + adversarial critic. 96.7% adversarial detection. 92% hallucination detection. The critic catches what consensus misses.

🔬

The Critic Effect

In 23% of cases where majority vote accepted a hallucination, the dedicated critic identified the error through structural analysis — checking reasoning chains, not counting votes. Full results →

Security Application

Multi-model divergence as prompt injection detection.

🛡️

AI Security Auditing

PoT identified critical vulnerabilities in major AI frameworks that single-model review missed — including prompt injection bypasses, RCE vectors, and compliance risks. Details published after responsible disclosure windows.

🔍

Injection Detection

Different models have different vulnerability profiles. When the same input produces divergent responses, that's a reliable injection signal. The more models, the harder coordinated attacks become.

📋

Forensic Traceability

Every deliberation is recorded: which model complied with an injection, which resisted, what the critic flagged. Epistemic Blocks create a growing corpus of attack signatures.

EU AI Act Compliance

Verifiable audit trails for high-risk AI systems. Regulation (EU) 2024/1689.

📋

Art. 9 — Risk Management

High-risk AI systems require continuous risk assessment with documented evidence. Epistemic blocks capture what was assessed, by which models, where disagreement occurred, and what confidence level was reached — creating the audit trail Art. 9(2) demands.

🔍

Art. 13 — Transparency

Users must be able to interpret AI outputs and use them appropriately. ThoughtProof's multi-perspective verification with explicit confidence scoring and dissent preservation implements transparency by design — not as an afterthought.

👤

Art. 14 — Human Oversight

The Act requires humans to "fully understand the capacities and limitations" of high-risk AI. Epistemic blocks present structured disagreement rather than false consensus — enabling the informed oversight Art. 14(4) requires.

Art. 43 — Conformity Assessment

Conformity assessment requires documented evidence across all Chapter 2 requirements. A chain of epistemic blocks provides verifiable, timestamped, provider-neutral records that auditors and notified bodies can inspect.

🔐

Provider-Neutral by Necessity

Conformity assessment cannot rely on tools built by the AI provider being assessed. ThoughtProof is structurally independent — no vendor can grade its own homework. The protocol works across all model providers.

🏛️

Open & Auditable

Regulators and notified bodies can inspect the verification methodology end-to-end. No black boxes. Every block, every critique, every confidence score is traceable and reproducible.

Built for High-Risk Domains

Where AI decisions have real consequences, single-model confidence isn't enough.

🏥

Healthcare AI

AI agents making clinical recommendations, triaging patients, or summarizing records need verifiable reasoning — not just confidence scores. Epistemic blocks document why a recommendation was made, which models disagreed, and where a physician should look twice. Built by a healthcare practitioner.

💰

Autonomous Finance

AI agents with wallet access can transfer funds based on corrupted reasoning. ThoughtProof verifies the decision chain before execution — catching prompt injection, hallucinated instructions, and social engineering that single-model systems miss. @pot-sdk2/pay →

⚖️

Legal & Compliance

When AI generates legal analysis, contract reviews, or regulatory assessments, the reasoning must be auditable. Multi-model verification with dissent preservation creates the evidence trail that legal teams and regulators expect.

Data Residency & Deployment

Zero data leaves your infrastructure. Fully self-hostable. No ThoughtProof servers involved.

🏠

On-Premise

Run with Ollama, vLLM, or any OpenAI-compatible endpoint. Models and data stay on your hardware. Full air-gap capable.

docker run thoughtproof/pot-cli \
  --provider http://ollama:11434
🔑

BYOK — Bring Your Own Keys

Use any LLM provider with your own API keys. ThoughtProof never sees your credentials, prompts, or outputs. No intermediary servers.

🇪🇺

GDPR & EU Data Sovereignty

No data processing by ThoughtProof means no GDPR data processing agreement needed. You control where your data lives — EU, on-prem, or air-gapped.

Build With ThoughtProof

From CLI to SDK to Agent-to-Agent trust infrastructure.

📦

pot-sdk

JavaScript/TypeScript SDK. Drop-in verification for any AI pipeline. verify() any output in one line. Plugin architecture: friend (persistent critic), graph (structural verification), bridge (agent-to-agent trust). npm →

🤝

A2A Trust Protocol

Agent-to-agent verification via MCP. Agents verify each other's reasoning chains before acting on outputs. Transitive trust with configurable decay. The missing trust layer for multi-agent systems.

🖥️

pot-cli

Run epistemic blocks from the terminal. Connect any model. Audit any output. npm install -g pot-cli

Protocol Resources

Read the specification. Review the code. Verify everything.

Read the Specification pot-sdk on npm pot-sdk on npm pot-cli on npm

npm install pot-sdk  ·  npm install -g pot-cli

Contact: raul@thoughtproof.ai