Open Source Reliability Harness: Make your agents follow rules. One line of code to enforce, trace, and improve.
-
Updated
May 14, 2026 - Python
Open Source Reliability Harness: Make your agents follow rules. One line of code to enforce, trace, and improve.
Static linter for AI agent configs, tool descriptions, and system prompts with zero-LLM CI gating
Production operations framework for AI-powered SaaS. The architectural patterns, failure modes, and operational playbooks that determine whether your AI systems scale profitably or fail expensively.
The flight recorder for AI agents. Tamper-proof, ISO 42001 + HIPAA-compliant audit trails for everything AI agents do on your system. File changes + shell commands + PHI detection + agent identity.
Cryptographic audit receipts for AI coding agents. Ed25519 + Merkle + RFC 3161 TSA. Supports Claude Code & Cursor.
PolicyBind is an AI Policy-as-Code platform that enables organizations to govern AI usage through a unified model registry, real-time token-based access control, and automated compliance reporting for frameworks like the EU AI Act and NIST AI RMF.
Git for AI workflows — Immutable, versioned audit trails as books and chapters. Web UI, CLI, API & agent-ready.ks and chapters. Web UI, CLI, API & agent-ready.
ai-agents llm-ai gpt-workflows ai-audit execution-logging ai-research reproducibility observability provenance ai-operations research-infrastructure operational-memory
Append-only event kernel with Ed25519-signed Merkle checkpoints. Every AI action gets a verifiable receipt.
RIO Receipt Protocol — Cryptographic proof for AI actions. Open standard for tamper-evident receipts, hash-chained ledgers, and independent verification.
A visual learning library for ISO/IEC 42001:2023 - the international standard for AI Management Systems. Reference cards, memory cards and deep dives.
A cool lil Hallucination detector for GenAI tools ! Hope you like it <3
EU AI Act compliance logging standard — structured one-line codes mapping AI reasoning (value, evidence, source hierarchies) to Traceability, Record-keeping, and Incident Reporting obligations under Articles 12, 13, 14, 15, and 73.
The Rego policy library for AI compliance: 94 policies covering the EU AI Act, NIST AI RMF, FAA/EASA/ICAO aviation, FERPA, fair lending, and 10+ more frameworks.
This repository defines a reproducible Layer-0 functional compliance specification for Large Language Models.
A tool for auditing bias through large language models
Hardened Public Release of KAIROS invocation governance framework. Includes invocation terms, ethical compliance clauses, regulatory mapping, and sample outputs. Licensed under CC BY-NC-ND 4.0. License: Do not auto-generate via GitHub. Use hardened License.txt
Sheldon K. Salmon — AI Reliability Architect. Creator of the AION Constitutional Stack and the CERTUS certainty‑engineering methodology. He designed, directed, and red‑teamed VERITAS — applying epistemic scoring, Uncertainty Mass, and permanent STP seals to community crisis data. Code is open source. The judgment is not.
Compliance-as-code for AI systems: evaluate AI apps against EU AI Act, NIST AI RMF, and OPA/Rego policies.
Cryptographically verifiable audit trail for AI systems — Ed25519 signed, hash-chained Decision Receipts with ISO 42001 / NIST AI RMF compliance. No blockchain, no SaaS.
Add a description, image, and links to the ai-audit topic page so that developers can more easily learn about it.
To associate your repository with the ai-audit topic, visit your repo's landing page and select "manage topics."