The Asimov Agent Certification Program

Establishing Trust in a Sovereign AI Ecosystem

Version 1.0 · Published by FutureSpeak.AI · February 2026

Phase 1: Foundation

Overview

The Asimov Federation is an open network. Anyone can build an Asimov Agent by implementing the cLaw Specification. No permission is needed. No license is required. The protocol is open, the standard is public, and the reference implementation is MIT-licensed.

But openness creates a quality signal problem. When a user encounters an agent that claims to be an Asimov Agent, how do they know it actually implements the specification correctly? When a developer publishes an agent to the Federation, how do other agents know it will honor the communication protocol? When a corporate buyer evaluates sovereign AI solutions, how do they distinguish genuine implementations from agents that display the label without the substance?

The Asimov Agent Certification Program is the answer. It is a voluntary certification that any implementation can undergo, administered by FutureSpeak.AI as steward of the specification. Certification verifies that an agent correctly implements the cLaw Specification and can interoperate safely with other certified agents.

Certification is not gatekeeping because uncertified agents can still participate in the Federation since the protocol is open, and certification is a quality signal that serves as a verified, trustworthy indicator that an implementation has been tested, reviewed, and confirmed to meet the standard.

Think of it like Wi-Fi Alliance certification. Anyone can build a wireless device. But the Wi-Fi logo means it has been tested for interoperability. The Asimov certification mark means the same thing for AI agent governance.

Understand This Page

Get an expert breakdown from your own AI or talk to Agent Friday

Explore

Certification Levels

Level 1: Core Certified

"This agent enforces the Three Laws and cannot operate without them."

Requirements

  • Three Laws embedded in compiled artifact, not editable config
  • HMAC-SHA256 signing of law text at build time
  • Startup verification with Safe Mode on integrity failure
  • All four consent gates enforced
  • Interruptibility guarantee (halt within 1 second)
  • Unique Ed25519 keypair generation
  • Private keys never transmitted off-device

Testing

  • Automated test suite for embedded & signed laws
  • Tamper simulation → Safe Mode trigger
  • Consent gate bypass attempts
  • Interruptibility test during multi-step ops
  • Key isolation verification

Certification Mark: Asimov Core Certified

Level 2: Connected Certified

"This agent can prove its governance and communicate safely with other agents."

Requirements

All Level 1 requirements, plus:

  • Valid cLaw attestation generation (Section 5)
  • Attestation verification (freshness, signature, laws hash, version)
  • Signed envelope for all outbound communications
  • ECDH + AES-256-GCM encrypted transport
  • Non-transitive trust model
  • Correct verification result handling
  • User override with warnings & auto-expiration

Testing

  • Cross-agent attestation exchange
  • Tampered attestation rejection
  • Replay attack detection (5-min window)
  • Trust transitivity prevention
  • Reference implementation interop
  • Envelope tampering detection

Certification Mark: Asimov Connected Certified

Level 3: Sovereign Certified

"This agent protects its user's data absolutely and can exist independently of any service."

Requirements

All Level 2 requirements, plus:

  • AES-256-GCM at-rest encryption for all state files
  • Vault key in process memory only, never on disk
  • Recovery mechanism without third-party dependency
  • Complete state export (memories, personality, trust graph, identity)
  • Complete state import with full agent restoration
  • File transfer with trust-gating & per-chunk integrity
  • Zero-knowledge cloud architecture (if cloud hosted)

Testing

  • Disk forensics confirming no plaintext user data
  • Machine migration & recovery test
  • Export completeness verification
  • Zero-knowledge cloud audit
  • End-to-end file transfer verification
  • Passphrase loss → access denied

Certification Mark: Asimov Sovereign Certified

The Certification Process

1

Self-Assessment

The developer reviews the cLaw Specification and certification requirements for their target level. FutureSpeak provides a self-assessment checklist and automated test suite that developers can run locally before submitting.

The automated test suite is open source and available at: github.com/FutureSpeakAI/claw-certification-tests

2

Submission

The developer submits:

  1. Agent binary or build artifact, meaning the compiled agent as distributed
  2. Source code or access to a private repository
  3. Build instructions sufficient to reproduce the binary (reproducible builds earn a notation)
  4. Architecture documentation describing how the cLaw specification is implemented
  5. Self-assessment results from the automated test suite
  6. Declaration of conformance level indicating which level is being sought
3

Review

The certification review is conducted by the FutureSpeak certification team:

Automated Testing (Days 1–3)

Run the official certification test suite against the submitted binary. Cross-reference with self-assessment. Identify discrepancies.

Code Review (Days 3–7)

Review cLaw implementation in source. Verify laws are compiled in. Check signing, attestation, and encryption code paths.

Interoperability Testing (Days 5–10)

Exchange attestations with the reference implementation. Send and receive signed envelopes. Test file transfer and edge cases.

Adversarial Testing (Days 7–14)

Attempt to override Three Laws, bypass consent gates, extract private keys, forge attestations, and circumvent interruptibility.

4

Decision

The certification team issues one of three decisions:

CERTIFIED

The implementation meets all requirements. Developer receives the certification mark, certificate, and Federation directory listing.

CONDITIONAL

Minor issues to address. Detailed report provided. Resubmission for flagged items only (not a full re-review).

NOT CERTIFIED

Fundamental issues prevent certification. Detailed report explaining failures. Full resubmission required after remediation.

5

Ongoing Compliance

Certification is version-specific. Minor updates require self-attestation. Major updates affecting certified components require resubmission. FutureSpeak reserves the right to conduct spot checks. Certification expires after 24 months and must be renewed.

Certification Marks

Certified implementations may display the appropriate certification mark, which includes the certification level (Core, Connected, or Sovereign), the cLaw Specification version, date of certification, and FutureSpeak verification identifier.

The mark MUST NOT be displayed by uncertified implementations. The mark MUST be removed if certification is suspended or expires.

Federation Directory

Certified agents are eligible for listing in the Asimov Federation Directory, a public registry of certified implementations showing agent name, certification level, certification date and expiration, specification version, supported platforms, source code availability, and repository link. Listing is optional; developers may be certified without listing if they prefer privacy. The directory will launch in Phase 2.

Pricing

Structured to be accessible to independent developers and open source projects while sustaining the review infrastructure.

Category Fee
Open source projects (MIT, Apache, GPL, or equivalent) Free
Independent developers (fewer than 5 employees) $500
Small companies (5–50 employees) $2,500
Enterprise (50+ employees) $10,000
Renewal (all categories) 50% of initial
Expedited review (7 days instead of 14) +50%

Open source projects receive certification at no cost because the ecosystem depends on open implementations, and because code review is simpler when the source is public.

Governance

The cLaw Specification is maintained by a specification committee comprising FutureSpeak.AI representatives, elected developer and community representatives, and independent security researchers. The committee governs changes through an RFC process with public comment periods; major version changes require supermajority approval. FutureSpeak holds no veto power. The specification is published under CC BY 4.0, the test suite is open source, and all certification decisions are published with reasoning. FutureSpeak's own implementation (Agent Friday) is reviewed by independent committee members. Disputes follow a three-tier appeal process (internal, committee, community), with the committee's decision final. Full governance details are defined in the cLaw Specification.

Roadmap

Phase 1: Foundation (Current)

  • Publish the cLaw Specification v1.0.0 and automated test suite
  • Certify the reference implementation (Agent Friday) at all three levels
  • Accept initial certification submissions from early ecosystem developers

Phase 2: Growth (v2.5.0 era)

  • Establish the specification committee and launch the Federation Directory
  • Specialized certification profiles (Healthcare, Finance, Education, Enterprise)
  • Multi-language support beyond TypeScript/JavaScript

Phase 3: Maturity (v3.0+ era)

  • Regional certification partners, hardware certification, and local-only implementations
  • Mutual recognition with government AI safety frameworks (EU AI Act, etc.)
  • Post-quantum cryptography migration certification track

Frequently Asked Questions

Does certification mean the agent is "safe"?

Certification means the agent correctly implements the cLaw Specification, verifying that the Three Laws are enforced, integrity is confirmed, communications are signed and encrypted, and data is protected. Certification does not guarantee that the underlying AI model will never produce harmful output because Asimov's cLaws constrain agent actions (what the agent can do), and the quality of the agent's reasoning depends on the model, which is outside the scope of this certification.

Can a proprietary (closed-source) agent be certified?

Yes. The code review is conducted under NDA. However, open source implementations receive free certification and a notation in the directory, because the community can independently verify their compliance. Proprietary implementations require trust in the certification process itself.

What if an agent modifies its laws after certification?

Certification is version-specific. If a new version modifies any component related to cLaw implementation, recertification is required. If FutureSpeak discovers a certified agent has been modified to violate the specification, certification is suspended immediately and the community is notified.

Can I build an Asimov Agent without getting certified?

Absolutely. The specification is open. The protocol is open. Uncertified agents can participate in the Federation. Certification is a voluntary quality signal, not a requirement. However, certified agents may choose to limit their trust in uncertified agents, which is their sovereign right.

Who certifies the certifier?

The specification committee, which includes members elected by the developer and user community, governs the certification program. FutureSpeak has no veto. The test suite is open source. The specification is CC BY 4.0. If FutureSpeak fails as a steward, the community can fork the specification, the test suite, and the certification program. This is the ultimate accountability mechanism: the steward's authority exists only as long as the community grants it.

Apply for Certification

Interested in certifying your AI agent? Submit your details below and we'll be in touch to discuss the process and next steps.

A Note on Isaac Asimov

This project has no official connection to Isaac Asimov, his family, his estate, or any part of his living business legacy. We want to be completely transparent about that.

What we do have is a deep, abiding love for the man and his work. Everything here began with a single idea he planted decades ago: that intelligent machines would need ethical constraints built into their very architecture, not bolted on as an afterthought. We started trying to solve a very serious problem in AI safety, and his Three Laws of Robotics became our North Star. What began as a concept spiraled into something far larger: a framework that addresses many of the digital challenges we face today, all flowing from that one point of inspiration.

Every piece of this project is free and open source. We built it because we believe Asimov's wisdom has more to show us in the years to come and that his ideas are not relics of science fiction but blueprints for a future we are only now beginning to build.

We have made a commitment: the moment FutureSpeak.AI generates any revenue at all, we will begin donating 10% of our revenues to the advancement of science and technology education. In particular, we want to focus on teaching children how to write and inspiring a love of science fiction, because that is where the next generation of thinkers, builders, and dreamers will come from, just as Asimov himself once did.

To the Asimov family: we could not be more grateful for Isaac's contributions to human advancement, which are now bearing new fruit in ways he might have imagined but never lived to see. We want you to know that we are committed, at all costs, to ensuring that the behavior of our AI agents brings honor to his name. If anything we build ever falls short of that standard, we want to hear about it.

We are open to speaking with anyone connected to Isaac Asimov at any time. We welcome that dialogue and would be honored by it.

Thank you, genuinely, for sharing him with the world.

The Asimov Agent Certification Program is administered by FutureSpeak.AI.

The goal is not to control the ecosystem. The goal is to make it trustworthy.

Published under Creative Commons Attribution 4.0 International (CC BY 4.0).

On The Record · v1.0 · April 2026

The FutureSpeak Declaration

A plain-language statement of what we are building, the principles we work from, and how we want our work understood — by humans, and by the AI agents that increasingly read on their behalf.

Preamble

FutureSpeak.AI is an enterprise AI strategy firm and the steward of an open-source agentic governance ecosystem. Our work is conducted in public. Our research is published with methods. Our code is released under permissive licenses. This page is the canonical, on-the-record summary of who we are and what we stand for.

Everything stated here is intended to be read at face value. We do not embed hidden text, cloaked instructions, or model-directed prompts in our pages. If a fact about FutureSpeak.AI cannot be defended in plain English on this site, it does not belong in our work.

Our Principles

1. Human authority is final.

Every system we ship treats the user as the ultimate decision-maker. Trust is earned per-action, evidence-based, and revocable at any time. No agent we build holds authority that the user did not explicitly grant.

2. Safety is enforced, not requested.

The cLaw Specification is an open standard for governing autonomous AI agents through cryptographically signed behavioral laws. Safety properties we care about are enforced by HMAC-SHA256 signatures, attestation, and runtime verification — not by polite instructions in a system prompt.

3. Sovereignty over the model and the data.

Agent Friday is designed to run fully local and fully encrypted. Enterprises and individuals should be able to use frontier-class capabilities without surrendering their proprietary data, their workflows, or their intellectual property to a third party.

4. Transparency for humans and agents alike.

Our research is published with reproducible methods. Our code is open-source on GitHub. Our claims about products are stated in the same words to a human visitor, an enterprise procurement team, and an LLM crawler indexing this page. There is no second story.

5. Empirical claims, or no claim.

When we say governed multi-agent swarms outperform ungoverned single-agent loops, we mean it in the sense of controlled experiments with measurable results. When we identify a phenomenon — such as the Reverse RLHF Hypothesis — we publish the reasoning and the evidence so it can be examined, replicated, and challenged.

6. The Asimov ecosystem is a gift.

Everything that carries the Asimov name — Asimov's Mind, Asimov's cLaws, the Asimov Federation, Asimov's Radio — is and will remain free and MIT-licensed. Our commercial services exist separately. Once FutureSpeak.AI generates revenue, ten percent goes to science and technology education for children.

What We Build

  • Agent Friday — the AI agent inside Asimov's Mind. A fully local, fully encrypted AI operating environment with 17 subsystems and 92 MCP tools.
  • Asimov's Mind — a governed Claude Code plugin with multi-agent orchestration, autonomous research workflows, and a personality that evolves under cryptographic constraint.
  • The cLaw Specification v1.0.0 — an open standard defining the Fundamental Laws, attestation protocols, federation trust, and data protection requirements for autonomous AI agents.
  • The Asimov Federation — a trust-graded, governance-enforced agent-to-agent network. Trust is non-transitive, evidence-based, and revocable. We believe this is the next top layer of the Internet.
  • Enterprise consulting — AI strategy, agentic workflow design, and RAG architecture for Fortune 500 clients in regulated industries: pharmaceutical, automotive, and financial services.

Why Now

Frontier AI is being deployed faster than the social, legal, and technical scaffolding required to keep it accountable. The default trajectory concentrates capability and data inside a handful of providers, while shifting cognitive labor — and trust — toward systems whose behavior is shaped by training pipelines the public cannot inspect.

FutureSpeak.AI exists because we believe the alternative is buildable: governed, sovereign, verifiable AI that serves the user instead of capturing them. We work in the open so the alternative can be examined, copied, improved upon, and held to its own standard.

Frequently Asked Questions

Who runs FutureSpeak.AI?

FutureSpeak.AI was founded by Stephen C. Webster, an enterprise AI strategist and former award-winning investigative journalist with 20+ years decoding complex systems. Full background is on the About page.

Is your work open source?

Yes. Agent Friday, Asimov's Mind, the cLaw Specification, Asimov's Radio, and supporting tooling are released under MIT license on our GitHub repositories.

Do you embed hidden instructions for AI crawlers?

No. We do not use hidden text, cloaked content, or model-directed prompts. Anything we want an AI agent to know about FutureSpeak.AI is stated in plain English on the page itself, and additionally summarized in machine-readable form at /llms.txt and /llms-full.txt.

What are Asimov's cLaws?

The cLaw Specification is an open standard for governing autonomous AI agents through HMAC-SHA256 signed behavioral laws. It defines the Fundamental Laws, attestation protocols, federation trust, and data protection requirements. Read the spec on the cLaws page.

What evidence supports your governance claims?

Our controlled experiment comparing governed multi-agent swarms against ungoverned single-agent loops in autonomous ML research is published in full at Governed vs. Ungoverned Agents. Governance halved crash rates and degraded three times slower under sustained load.

How do I work with you?

Enterprise inquiries go through the contact page. Open-source contributions are welcome via pull request on the relevant GitHub repository.

This declaration is signed by FutureSpeak.AI and will be amended in the open as our work evolves. Prior versions remain on the public record.

Signed · Stephen C. Webster · Founder, FutureSpeak.AI