HomeVisionPrinciplesGet StartedGitHub
Background Image
Enrich or Extract

AI that doesn't serve humanity is extracting from it.

If you can't audit the ethics, they're marketing. Here are six requirements for verifiably ethical AI — and why closed-source systems can't meet them.

Try it yourself. Everything on this page is implemented today.

Free to install · No signup required (unless using our privacy-protecting LLM proxy)

CIRIS isn't productivity AI. It's runtime governance for agentic AI — infrastructure for high-stakes deployment where misalignment kills.

The only open stack enforcing all six ethical requirements at runtime (as of Dec 2025). Open an issue if we've missed a peer.

The Six Requirements

All six. Not five.

An AI system only qualifies as verifiably ethical if it meets ALL six. Governance frameworks and safety filters are useful — but they're layers around agents, not ethical agents themselves.

1. Published Principles

The agent must be bound to a public ethical framework: Beneficence, Non-maleficence, Integrity, Transparency, Autonomy, and Justice. Not guidelines. A formal document the agent is obligated to follow.

2. Runtime Conscience

Every action passes through ethical checks before execution. Not a post-hoc filter — part of the decision loop itself.

3. Human Deferral

When uncertain or facing potential harm, the agent defers to humans with full context. Built into the workflow, not a suggestion.

4. Cryptographic Audit

Every action and rationale recorded in an immutable, signed ledger. Not 'we log some things.' Everything. Trace exactly why the agent did what it did.

5. Bilateral Consent

Consent goes both ways. Humans can refuse data access. The agent can refuse requests that violate its principles. Neither party compromises.

6. Open Source (AGPL-3.0)

Ethical AI cannot be closed source. You can't audit what you can't see. 'Trust us, it's ethical' is not ethical. Show the code.

The Research Agrees

Peer-reviewed papers, regulatory bodies, and transparency indices document the same gaps.

The Runtime Gap

Design-time ethics don't execute.

Most ethical AI stops at governance frameworks. CIRIS provides runtime governance — enforcing principles during operation, not just at design time. 'Runtime verification mechanisms ensure principles remain strictly adhered to during operation.' — Springer, AI and Ethics (2025)

Transparency Crisis

No company passes.

'No foundation model developer gets a passing score on transparency. None score above 60%.' — Stanford Foundation Model Transparency Index. Closed source means structural opacity.

Ethics Washing

A peer-reviewed term.

'The practice of signaling commitment to ethics without genuinely putting it into practice.' — Carnegie Council. If your ethical AI is a press release, not runtime code, researchers have a name for that.

Guardrails ≠ Conscience

Safety tools solve a different problem.

LlamaFirewall

Meta's guardrail framework mitigates prompt injection and insecure code. Security guardrails — not ethical conscience.

NeMo Guardrails

NVIDIA's system addresses adversarial prompting and injection attacks. Runtime safety — not reasoning about values.

The Distinction

Safety guardrails block bad outputs. Ethical conscience reasons about values. Safety prevents harm. Ethics reasons about right and wrong. Different problems.

EU AI Act Article 14

CIRIS implements regulatory requirements.

The EU AI Act mandates human oversight for high-risk systems. CIRIS's deferral mechanism implements Human-in-Command, Human-in-the-Loop, and Human-on-the-Loop. Most AI systems don't have a deferral mechanism at all.

Common Objections

And why they're wrong.

Constitutional AI

A training technique (RLHF variant). Shapes behavior during training. Does not enforce ethics at runtime. Training is not architecture.

Ethics Boards

Governance, not architecture. Ethics boards review policies. They don't gate every action. CIRIS enforces on every action. Documents don't execute.

Safe Models

Safety prevents harmful outputs. Ethics reasons about values. A 'safe' model can still make unethical decisions. Different problems. Both matter.

The Current Landscape

As of December 2025, we haven't found another system meeting all six.

Based on publicly available documentation as of December 2025. If we've missed something or gotten something wrong, open an issue.

ProjectRuntime SystemPrinciplesConscienceAudit TrailConsentAGPL-3.0
CIRISYesYesYesYesYesYes
MI9 FrameworkPaper onlyNoConceptConceptNoNo
HADA ArchitecturePoC onlyNoNoLoggingNoNo
Superego PrototypeResearchPartialPartialPartialNoNo
METR (nonprofit)Evaluation onlyNoNoNoNoNo
Agentic AI FoundationStandards onlyNoNoNoNoNo
Manus AIYesNoNoLimitedNoNo

Sources: arXiv (MI9, HADA, Superego), Wikipedia (METR, Manus AI), WIRED (Agentic AI Foundation)

The AGI Question

Decentralized alignment beats centralized control.

Many Aligned Agents

Not one unaligned god.

The dominant AI safety narrative assumes one superintelligent system that must be perfectly aligned or humanity loses. CIRIS rejects that frame. Instead: many smaller agents, each bound to published principles, each auditable, each deferring to human authority. Distributed governance, not concentrated power. No single point of failure. No race to build God.

Distributed Governance

Power stays distributed. Each CIRIS instance answers to its local Wise Authority, not a central controller. Geopolitical risk from AI concentration is structural — the fix is architectural.

Aligned Baby-AGIs

Small, verifiable agents scaling horizontally. Each bound to principles. Each auditable. Each killable. The alternative to racing toward uncontrollable ASI is building many controllable agents that stay aligned.

No Single Chokepoint

Centralized mega-AGI means winner-take-all dynamics and single points of catastrophic failure. Decentralized aligned agents mean no one entity controls the stack. Humanity keeps the keys.

MI9 Framework

A research architecture proposing runtime governance for agentic AI. Theoretical framework only — no deployed system, no published principles, no cryptographic audit. Paper, not product.

HADA Architecture

Reference architecture wrapping agents with stakeholder roles (ethics, audit, customer). A proof-of-concept demo, not a general-purpose ethical agent platform. Research, not runtime.

Manus AI

A deployed autonomous agent — but not alignment-focused. No published principles, no ethical reasoning layer, no deferral mechanism, no cryptographic audit, no consent framework. Capable, but not verifiably aligned.

The Agentic AI Foundation

OpenAI, Anthropic, Block — building standards, not ethics.

The companies that could build ethical agentic AI are instead building agent communication protocols. Useful work. But it doesn't address conscience, principles, consent, or audit. They're standardizing how agents talk — not how agents reason about right and wrong.

What You'll Experience When You Install

Transparent Reasoning

Watch the agent's ethical checks in real-time. See why it chooses each action.

Principle-Checked Answers

Every response passes through conscience validation against the published ethical principles.

Deferral in Edge Cases

When uncertain, the agent asks you instead of guessing. Human oversight built into the loop.

Deploy for safety-critical use cases: content moderation, crisis response, regulatory compliance, AI governance research.

Verify It Yourself.

Install it. Audit the code. Join the auditors building uncompromisable AI.

Free to install · No signup unless using our LLM proxy · Your data stays on your device

Verify It Yourself.

pip install ciris-agent

The only open stack enforcing all six requirements end-to-end, in code, running in production. Audit it. Deploy it for safety-critical use cases: moderation, crisis response, governance. Tell us what's missing.