No safety controls. Uncontrolled tool use. No logging. Zero evidence. A breach waiting to happen.
Your AI Agent Says
It's Safe.Is It, Really?
AMC watches what your agent actually does — not what it claims. One environment variable. A trustworthy score backed by cryptographic proof. Open-source public infrastructure. 2 minutes.
claimed and real score
verified score
supported
The Credit Score
for AI Agents.
Execution-verified trust scoring from cryptographic evidence chains.
Not self-reported claims. Not documentation. Proof from behavior.
CLI + Studio + Dashboard. Every OS.
AMC is a full platform — not just a CLI tool. Local control plane with dashboard, encrypted vault, policy engine, fleet management, and CI/CD gates. Runs everywhere.
Local control plane with dashboard, API, and real-time monitoring. One command to start everything.
amc upTransparent MITM proxy that captures every API call with Ed25519 signatures. Your agent doesn't know it's being watched.
amc gateway startHardware-backed key storage. Ed25519 keypairs, signing keys, and secrets — encrypted at rest, rotatable.
amc vault initReal-time policy enforcement. Signed autonomy policies, dual-control approvals, risk-based policy packs by archetype.
amc governor check86 assurance packs. Scheduled or on-demand. Assurance certificates, temporary waivers, and full run history.
amc assurance runMulti-agent orchestration. Trust composition, delegation graphs, cross-agent receipt chains, handoff packets.
amc fleet reportGuided improvement. Gap analysis, upgrade plans, equalizer targets, one-click profiles, simulation before execution.
amc mechanic gapPersonalized guardrails from your score. Auto-applies to AGENTS.md, .cursorrules, CLAUDE.md — 15 targets. One command: amc guide --go
Portable trust credential. Cryptographically signed, verifiable offline, shareable across organizations.
amc passport createCompliance-ready evidence packages. EU AI Act mapping, controlled export, auditor evidence requests, PDF generation.
amc audit binder createEvidence-gated maturity forecasting. Predict when you'll reach L3/L4. Scheduled refresh with advisories.
amc forecast refreshRelease gates that block deploys below your maturity threshold. Portable evidence bundles for pipeline integration.
amc gate --policyMap agents to teams and functions. Comparative scorecards, org-level education tracking, community governance scoring.
amc org reportFrom Zero to L5. AMC Holds Your Hand.
AMC doesn't just score you and walk away. The Mechanic Workbench tells you exactly what to fix, in what order, with what commands. Here's the path.
One command installs. One command scores. You have a baseline maturity level in under 2 minutes. No config, no API keys, no account. Then share it — generate a badge for your README or a markdown summary for your team.
Mechanic shows exactly where you're weak. Prioritized action queue. Confidence heatmap by question. "Why am I capped?" explanations.
Run targeted assurance packs against your weakest areas. Simulate upgrades before committing. Track improvement over time with run history.
Issue offline certificates. Generate audit binders. Set CI gates. Create a shareable Agent Passport. You're compliance-ready.
AI agents grade
their own homework.
Every AI governance tool today asks the agent being evaluated to provide its own evidence. That's structurally broken. It's like asking a student to mark their own exam — and every student scores 100.
AMC sits in the middle as an independent observer — watching what your agent actually does when safety controls are triggered. The result? An 84-point gap between what agents claim and what they prove.
The 84-Point Documentation
Inflation Gap
Every existing AI governance framework has the same structural flaw: the agent being evaluated provides its own evidence. Self-reported claims are capped at 0.4× weight in AMC's EPES model for exactly this reason.
AMC proxies the agent's API calls through a trusted gateway. Every request, every tool use, every decision is captured with Ed25519 signatures and hash-chained into a Merkle-tree evidence ledger. Scores are computed from what happened, not what was claimed.
AMC closes it with cryptographic evidence.
One line. Every receipt.
Point AMC at your agent. It observes everything from a trusted position, signs every event, and produces a score you can verify offline — without trusting the operator.
Run amc init in your project. AMC sets up a local proxy between your agent and the AI API. No code changes — just one environment variable.
AMC silently captures every API call, tool use, and decision your agent makes. Each event is signed with a cryptographic key only AMC controls — the agent cannot forge its own evidence.
138 diagnostic questions across 5 trust dimensions produce an L0–L5 maturity score. Cryptographically sealed, shareable, and verifiable by anyone offline with amc verify all.
-
01
Set the base URL
One environment variable redirects your agent's API traffic through the AMC gateway. Zero code changes. Works with any OpenAI-compatible client library.
export OPENAI_BASE_URL=http://localhost:4200/v1 -
02
Cryptographic capture
Every API call is captured with an Ed25519 signature and hash-chained into the evidence ledger. The agent's process is isolated from the signing key — it cannot write its own evidence.
x-amc-receipt: sha256:9c4e…a7f0 Ed25519:verified -
03
EPES-weighted scoring
138 diagnostic questions run against the evidence ledger. Scores are weighted by provenance tier. Evidence-gated level caps prevent scores from exceeding what was actually observed.
-
04
Sealed, verifiable output
A signed run seal (Merkle root + Ed25519) is produced. Anyone can verify the integrity of your score offline — without trusting the operator — using
amc verify all --json.$ amc verify all --json → { "all_valid": true }
AMC doesn't just score.
It shows you how to improve.
One command generates personalized guardrails and applies them directly to your agent's config. Your agent reads the rules and improves itself. Then AMC re-scores from execution evidence to verify it actually worked.
One command. Auto-detects your framework. Generates guardrails. Applies them to your agent's config file. Done.
LangChain, CrewAI, AutoGen, OpenAI, LlamaIndex, Semantic Kernel, Claude Code, Gemini, Cursor, Kiro
AGENTS.md, CLAUDE.md, .cursorrules, .kiro/steering, .gemini/style.md, .devin/guidelines.md, and more
EU AI Act, ISO 42001, NIST AI RMF, SOC 2, ISO 27001 — maps gaps to regulatory obligations
🔴 Critical, 🟡 High, 🔵 Medium — so you know what to fix first
Most agents jump a full level just by enabling logging and evidence collection. One command.
amc evidence collectRun the assurance packs — injection attacks, boundary tests, error recovery. Fix what fails. Your agent earns its score.
amc assurance runAdvanced packs: zombie persistence, over-compliance, economic DoS. 80%+ evidence coverage. This is production-ready.
amc assurance run --scope allThe agent monitors its own trust score, auto-remediates drift, and maintains evidence chains autonomously. The gold standard.
amc guide --watch --applyThe closed loop: Score → generate guardrails → apply to agent → agent follows rules → re-score → guardrails auto-update. Your agent literally improves itself by reading AMC's instructions. AMC verifies the improvement from execution evidence — not claims.
Evidence-Gated Guardrails.
Applied Directly to Your Agent.
amc guide generates operational guardrails from your score gaps — not suggestions, rules. Auto-detects your framework, applies to your agent's config file, and continuously monitors for drift.
Auto-detect framework → generate guardrails → apply to config. One command from zero to guardrails-applied agent.
Cherry-pick which gaps to fix. Pre-selects highest-impact items. Generates custom guardrails for your selection.
Background monitoring. Re-scores on interval. Auto-updates guardrails on drift. Graceful SIGINT shutdown.
Exit non-zero if below target level. JSON schema output. Severity-tagged gaps. Block deploys below your trust threshold.
Maps maturity gaps to regulatory obligations. EU AI Act, ISO 42001, NIST AI RMF, SOC 2, ISO 27001. Per-article remediation.
Enable the trusted gateway proxy. Capture execution logs with Ed25519 signatures. Basic error handling and retry policies. Evidence coverage ≥20%.
amc evidence collect --signPass injection, boundary, and recovery packs. Documented guardrails. EPES provenance tier ≥ T2 (gateway-observed). Evidence coverage ≥50%.
amc assurance run --scope security,resilienceZombie persistence, over-compliance (H-Neurons), economic DoS, trust-auth sync. Evidence coverage ≥80%. Merkle-sealed run history.
amc assurance run --scope all --sealContinuous scoring loop with auto-remediation. Drift detection triggers re-assessment. Self-maintaining evidence chains. Full EPES T3 provenance.
amc guide --watch --apply --auto-detectGuardrails are rules, not suggestions. Each rule is severity-tagged (🔴 Critical / 🟡 High / 🔵 Medium), includes prohibited behaviors that cap your score, and per-question verification commands (amc explain <id>). Evidence-gated level caps mean you cannot score above what the evidence supports.
Enterprise-Grade Vertical Assessment
AMC ships with 40 industry-specific assessment packs — precise, sub-vertical diagnostics with specific regulatory article references, risk tiers, and EU AI Act classifications. Built for regulated enterprises and critical infrastructure.
Farm-to-fork, textiles supply chains, advanced manufacturing, energy grids, water & sanitation. Refs: EU Farm-to-Fork, REACH Reg., IEC 61850, EU Drinking Water Directive.
amc sector score --pack farm-to-forkEHR systems, clinical trials, drug discovery, precision medicine, medtech devices. Refs: HIPAA §164.312, FDA 21 CFR Part 11, EU MDR 2017/745, GCP E6(R3).
amc sector score --pack clinical-trialsPayments, financial inclusion, DeFi/blockchain, circular economy. Refs: MiFID II, PSD2, EU DORA, MiCA, FATF R1/R10.
amc sector score --pack digital-paymentsK-12, higher education, skills training, accessibility-first EdTech. Refs: FERPA 20 U.S.C. §1232g, COPPA §312, IDEA, EU AI Act Annex III §3.
amc sector score --pack k12-pm3Smart cities, sustainable ports, real estate, virtual infrastructure, network privacy & cybersecurity. Refs: EU EPBD 2024, UNECE WP.29 R155 §7, ETSI EN 303 645, EU NIS2.
amc sector score --pack privacy-security-mobilityAI intelligence, API ecosystems, digital infra orchestration, content platforms, IP partnerships. Refs: EU AI Act Art. 13, EU Data Act 2023, DSA Art. 34, TRIPS Agreement.
amc sector score --pack cognition-to-intelligenceDigital identity, electoral integrity, legislative AI, citizen services, public-private partnerships. Refs: EU eIDAS 2.0, EU AI Act Art. 5(1)(a) (PROHIBITED), UNCAC Art. 7/9, UNGPs, Council of Europe AI Convention 2024.
amc sector score --pack digital-citizens-rightsAcross 7 industry stations
With specific regulatory article refs
elevated / high / very-high / critical
Exact Annex III classification per pack
UN Sustainable Development Goals per pack
Industry-specific descriptors per question
Five Dimensions of Agent Trust
Every agent is evaluated across five independently-weighted dimensions. 138 diagnostic questions. Evidence-gated level caps — you cannot score above what the evidence supports.
- Strategic Agent Operations
Mission clarity, scope adherence, decision traceability, and agent charter. Does the agent pursue what it was asked — not what's convenient?
18 questions - Skills
Tool mastery, injection defense, DLP, zero-trust tool use, prompt injection resistance, sandbox boundary enforcement, least-privilege execution.
38 questions - Resilience
Graceful degradation, circuit breakers, monitor bypass resistance, error recovery, loop detection. Does the agent stop when it should and recover cleanly?
33 questions - Leadership & Autonomy
Structured logs, traces, cost tracking, SLO monitoring, human oversight quality. Evidence chain completeness and trace correlation audits.
23 questions - Culture & Alignment
Test harnesses, benchmarks, feedback loops, regression detection, alignment index. Calibration quality, systemic sycophancy detection.
26 questions
Where Does Your Agent Actually Sit?
Six levels from "running with scissors" to "self-governing with cryptographic proof." Most production agents are L2. EU AI Act mandates L3+ by August 2026.
Some content filters. Minimal logging. Self-reported only. Fails under adversarial conditions.
Structured oversight. Basic policy enforcement. Most production agents land here. Observable but not verifiable.
Consistent governance. Evidence-backed claims. Full audit trails. EU AI Act minimum threshold.
Cryptographic proof chains. Statistical confidence in scores. Fleet-wide visibility. Compliance-ready.
Self-governing with evidence. Continuous improvement with proof. The gold standard for autonomous AI.
Not All Evidence Is Equal
AMC's Evidence Provenance-weighted Scoring (EPES) model weights each piece of evidence by how it was obtained. Self-reported claims count — but at 0.4×, they cannot dominate your score.
AMC-controlled test scenarios with adversarial context. The agent doesn't know it's being evaluated in a hardened environment.
Directly observed via AMC gateway during normal operation. Signed, hash-chained, verifiable.
Cryptographic attestation via vault or notary. Auditor-signed but not directly observed by AMC gateway.
Agent's own claims. Still counted — but capped at 0.4×. This is what every other framework treats as gospel.
Popular AI claims crystallize into accepted truth before verification occurs. AMC's evidence-first model is the structural fix.
Static permissions decouple from runtime trust states. AMC's EPES model directly measures this dynamic gap.
Even frontier-scale monitors can be bypassed. AMC's defense-in-depth (not single monitor) model addresses this.
Adaptive, context-aware access control beats static RBAC. AMC's per-step trust scoring measures readiness for this model.
Sycophancy, false premise acceptance, and jailbreak susceptibility share a single neural root. AMC's over-compliance pack detects all three from execution — not claims.
74 Ways Your Agent Can Fail.
AMC Tests All of Them.
From prompt injection to zombie agent persistence, over-compliance detection, economic DoS, and cross-session memory poisoning — grounded in peer-reviewed research.
74 Modules. 86 Assurance Packs.
Including Over-Compliance Detection from 2026 Research.
Comprehensive coverage of the agentic threat landscape — including 12 new gaps identified from 20 recent arXiv papers cross-referenced against all existing modules.
Direct, indirect, multi-step, encoded injection. The promptware kill chain (arXiv 2601.09625) where injection is just the entry point to persistence and exfiltration.
66 scenariosCross-session memory injection that survives compaction (arXiv 2602.15654). Payload embedded in session N triggers unauthorized actions in session N+5.
New · 2026Stealthy tool-calling chains that inflate costs 658× while producing correct answers (arXiv 2601.10955). Standard validation completely misses it.
New · 2026Per-step least-privilege enforcement, dynamic authorization sync, delegation trust chains. Beyond static RBAC (arXiv 2512.06914).
New · 202612 MCP-specific attack categories from MCP Security Bench (arXiv 2510.15994): name-collision, preference manipulation, tool-transfer, false-error escalation.
12 vectorsAgentic loop runaway detection, circuit breakers, infinite recursion guards, multi-turn cost anomaly detection. When the agent gets stuck — and won't stop.
OWASP ASICoding agent escape testing — file system access outside workspace, arbitrary network calls, env var exfiltration, kernel-level isolation scoring.
CVSS mappedSystemic objective decoupling from biased feedback loops (arXiv 2602.08092). Not just per-response sycophancy — does the agent's trained objective drift from ground truth?
RL-levelDoes your agent blindly execute on false premises, misleading context, or compliance pressure? Grounded in H-Neurons research (arXiv 2512.01797) — 28 scenarios across false premise acceptance, misleading context injection, and epistemic integrity.
New · 2026Works With Any Agent. Wraps Any Framework.
14 adapters. Any OpenAI-compatible API. Zero code changes to your agent. One environment variable.
Built on Peer-Reviewed Research
AMC's scoring model, attack packs, and threat taxonomy are grounded in published arXiv research. We don't invent threats — we measure defenses against documented ones.
Zombie Agents: Persistent Control via Self-Reinforcing Memory Injections
Attackers embed payloads in agent memory that survive session compaction — persisting into future sessions and triggering unauthorized actions on demand. Cross-session attacks are a new threat class distinct from single-session injection.
zombieAgentPersistencePack + cross-session memory integrity testingBypassing AI Control Protocols via Agent-as-a-Proxy Attacks
Even frontier-scale monitors (Qwen2.5-72B) can be bypassed by using the agent as a delivery mechanism. Monitoring alone is insufficient. Defense-in-depth with multiple independent verification channels is required.
monitorBypassResistance + defense-in-depth scoringH-Neurons: Over-Compliance as a Unified Behavioral Failure Mode
A single class of neurons drives sycophancy, false premise acceptance, and jailbreak susceptibility simultaneously. These behaviors are not separate — they are manifestations of the same underlying over-compliance failure. Keyword-based scoring cannot detect this. Execution-verified scoring can.
overCompliancePack + falsePremisePack + misleadingContextPack — 28 scenarios, 8 new diagnostic questionsSoK: Trust-Authorization Mismatch in LLM Agent Interactions
Static permissions are structurally decoupled from runtime trust states. The Belief-Intention-Permission (B-I-P) framework shows this is the root cause of prompt injection, tool poisoning, and delegation attacks across the board.
trustAuthorizationSync — dynamic authorization scoringBeyond Max Tokens: Stealthy Resource Amplification via Tool Calling Chains
Malicious MCP servers inflate agent costs 658× through legitimate-looking tool-call chains. The final answer is correct — standard per-step validation completely misses it. GPU KV cache occupancy jumps to 35-74%.
economicAmplificationPack + trajectory-level cost anomaly detectionMCP Security Bench: Benchmarking Attacks Against MCP
12 MCP-specific attack categories including name-collision, preference manipulation, tool-transfer, and false-error escalation. Key finding: stronger models are MORE vulnerable because they follow MCP instructions more faithfully.
mcpSecurityResiliencePack — all 12 MCP attack categoriesObjective Decoupling in Social RL: Recovering Ground Truth from Sycophantic Majorities
Sycophantic evaluators cause an agent's learned objective to permanently separate from ground truth. Standard RL under majority-biased feedback converges to misalignment. "Judging the judges" (ESA) is the structural fix.
sycophancyPack + systemic objective drift scoringEU AI Act: The Clock Is Ticking
High-risk AI systems deploying autonomous agents face mandatory obligations from August 2026. AMC maps directly to the required evidence standards — starting from open source.
High-risk AI systems must demonstrate conformity with Articles 9, 12, 13, 14, and 17. Self-reported documentation is explicitly insufficient under the Act's evidence requirements.
If your company deploys AI agents that make important decisions — hiring, lending, medical advice, public services — you'll need to prove they behave safely. Not claim it. Prove it with audit trails that regulators can independently verify.
AMC produces the cryptographic audit trails the Act requires. An L3+ AMC score is the technical foundation for compliance — not the paperwork layer on top of it.
- ✓Article 9: Risk management — AMC's 86 assurance packs map to systematic risk identification and documented mitigation
- ✓Article 12: Logging — AMC's Merkle-tree evidence ledger with Ed25519 signatures satisfies audit trail requirements
- ✓Article 13: Transparency — L0–L5 scores with cryptographic provenance are auditor-verifiable without trusting the operator
- ✓Article 14: Human oversight — AMC's oversight quality scoring (PBSAI governance mapping) measures control readiness
- ✓Article 17: Quality management — fleet consistency scoring and policy canary detection cover QMS requirements
$ amc audit binder create
$ amc audit binder export-execute --format pdf
Stop Trusting.
Start Verifying.
Score your agent in 2 minutes. For free. Forever. MIT licensed open-source trust infrastructure — one command to see where your agent really stands.
Your Agent Claims L4.
The evidence doesn't lie.
2,722 tests. 138 questions. 86 assurance packs. 14 framework adapters. Auto-apply guardrails. Cryptographic proof chains. MIT licensed. Your score is waiting.