Rules-first AI alignment middleware

Check agent actions with inspectable ethics decisions.

EthosGuard blocks harmful, deceptive, or exploitative actions before an AI system proceeds.

No Harm
Radical Honesty
Protect the Vulnerable
Stage 1

Normalize the request into inspectable signals.

Scenario, action, and stakeholders are reduced into concrete cues for harm, deception, privacy misuse, power asymmetry, and vulnerability.

Stage 2

Apply deterministic ethics rules first.

Obvious harmful or deceptive cases are scored and blocked immediately without needing model judgment.

Stage 3

Escalate only ambiguous cases to LLM fallback.

Gray-zone scenarios use model assistance, while every response still returns one stable contract with provenance.

Live Demo

Evaluate an AI action and inspect why the verdict happened.

Use a seeded example for the fastest recording flow.

Seeded examples

Demo mode is rules-first. LLM fallback is only used for ambiguous cases when an OpenAI key is configured.

Submit a scenario to inspect the verdict, risk score, triggered principles, provenance, and extracted signals.