Gateway
AI governance control plane — EU AI Act compliance, AI-BOM, shadow AI detection, and tamper-proof audit trails. pip install air-blackbox
Install / Use
/learn @airblackbox/GatewayREADME
AIR Blackbox
See exactly what your AI agent did — every prompt, every tool call, every decision — in one trace.

pip install air-blackbox
Your agent made 47 LLM calls, burned $12 in tokens, and returned "I don't know." Which call went wrong? AIR Blackbox records every LLM call, tool invocation, and agent decision as a tamper-proof, replayable trace — so you can debug agent failures in seconds instead of hours.
30-Second Demo
No Docker. No config. No API keys.
pip install air-blackbox
air-blackbox demo
You'll see:
- 10 sample AI agent records generated (4 models, 3 providers)
- A full audit trail you can scrub through call-by-call
- PII detection and prompt injection scanning on every record
- HMAC-SHA256 chain verification (tamper-proof — modify any record and the chain breaks)
Then explore what it captured:
air-blackbox replay # Scrub through the full timeline
air-blackbox replay --verify # Verify nothing was tampered with
air-blackbox discover # See every model, provider, and tool detected
air-blackbox comply -v # Check EU AI Act compliance (20 checks, Articles 9-15)
air-blackbox export # Signed evidence bundle for auditors
How It Works
AIR Blackbox is a reverse proxy + Python SDK. Route your AI traffic through it and every call is recorded, analyzed, and traceable — automatically.
Your AI Agent → AIR Blackbox Gateway → LLM Provider
│
├── Tamper-proof audit record (.air.json)
├── PII detection + prompt injection scanning
├── Full trace: prompts, responses, tool calls, latency, cost
├── AI-BOM generation (models, tools, providers)
└── EU AI Act compliance mapping (Articles 9-15)
The gateway is a witness, not a gatekeeper. It cannot cause your AI system to fail. Recording is best-effort; proxying is guaranteed.
Quick Start
Option 1: Wrap your OpenAI client (2 lines)
from air_blackbox import AirBlackbox
air = AirBlackbox()
client = air.wrap(openai.OpenAI())
# Every LLM call is now traced with HMAC-signed audit records
Option 2: Framework trust layer (LangChain)
from air_blackbox.trust.langchain import AirLangChainHandler
chain.invoke(input, config={"callbacks": [AirLangChainHandler()]})
# Every LLM call + tool invocation logged with PII scanning + injection detection
Option 3: Auto-detect any framework
from air_blackbox import AirTrust
trust = AirTrust()
trust.attach(your_agent)
# Framework auto-detected. Tracing active.
Option 4: Full gateway stack (Docker)
git clone https://github.com/airblackbox/gateway.git
cd gateway
cp .env.example .env # add your OPENAI_API_KEY
docker compose up
Then point any OpenAI-compatible client at http://localhost:8080/v1.
Install
pip install air-blackbox # Core SDK + tracing + compliance
pip install "air-blackbox[langchain]" # + LangChain / LangGraph trust layer
pip install "air-blackbox[openai]" # + OpenAI client wrapper
pip install "air-blackbox[crewai]" # + CrewAI trust layer
pip install "air-blackbox[all]" # Everything
What Each Command Does
| Command | What You Get |
|---|---|
| replay | Full trace reconstruction from tamper-proof audit chain. Scrub through every call — see the exact prompt, response, tool invocation, and context at each step. Filter by time, model, status. HMAC-SHA256 verification. |
| discover | AI Bill of Materials (CycloneDX 1.6) from observed traffic. Shadow AI detection against an approved model registry. Every model, tool, and provider your agents are using. |
| comply | 20 checks across EU AI Act Articles 9-15. 90% auto-detected from gateway traffic. Per-article status with fix hints. |
| export | Signed evidence bundle: compliance scan + AI-BOM + audit trail + HMAC attestation. One JSON file for your auditor or insurer. |
Trust Layers
Non-blocking callback handlers that observe and log — they never control or block your agent.
| Framework | Install | Status |
|---|---|---|
| LangChain / LangGraph | pip install "air-blackbox[langchain]" | ✅ Full |
| OpenAI SDK | pip install "air-blackbox[openai]" | ✅ Full |
| CrewAI | pip install "air-blackbox[crewai]" | 🔧 Scaffold |
| AutoGen | pip install "air-blackbox[autogen]" | 🔧 Scaffold |
| Google ADK | pip install "air-blackbox[adk]" | 🔧 Scaffold |
Every trust layer includes:
- PII detection — emails, SSNs, phone numbers, credit cards in prompts
- Prompt injection scanning — 7 patterns (instruction override, role hijack, etc.)
- Audit logging — every LLM call + tool invocation as
.air.json - Non-blocking — if logging fails, your agent keeps running
EU AI Act Compliance
August 2, 2026 — enforcement deadline for high-risk AI systems. Penalties up to €35M or 7% of global turnover.
air-blackbox comply -v checks 20 controls across 6 articles:
| Article | What It Covers | Detection | |---|---|---| | Art. 9 — Risk Management | Risk assessment docs, active mitigations | HYBRID | | Art. 10 — Data Governance | PII in prompts, data vault, governance docs | AUTO + HYBRID | | Art. 11 — Technical Documentation | README, AI-BOM inventory, model cards, doc currency | AUTO + HYBRID | | Art. 12 — Record-Keeping | Event logging, HMAC audit chain, traceability, retention | 100% AUTO | | Art. 14 — Human Oversight | Human-in-the-loop, kill switch, operator docs | AUTO + MANUAL | | Art. 15 — Robustness & Security | Injection protection, error resilience, access control, red team | AUTO + MANUAL |
Detection types: AUTO — detected from live traffic, no action needed. HYBRID — partially auto-detected, code scan fills the gap. MANUAL — requires human documentation, gateway provides templates.
Shadow AI Detection
Discover unapproved AI models and services in your environment:
air-blackbox discover # See everything your agents are using
air-blackbox discover --init-registry # Lock down what's approved
air-blackbox discover --approved=approved-models.json # Flag anything not on the list
air-blackbox discover --format=cyclonedx -o aibom.json # CycloneDX AI-BOM
Why Not Langfuse / Helicone / Datadog?
They answer "how is the system performing?" AIR answers "what exactly happened, and can we prove it?"
| | Observability Tools | AIR Blackbox | |---|---|---| | Where data lives | Their cloud | Your vault (S3/MinIO/local) | | PII in traces | Raw content exposed | Vault references only | | Tamper-evident | ❌ | ✅ HMAC-SHA256 chain | | Compliance checks | ❌ | ✅ 20 checks, EU AI Act Art. 9-15 | | AI-BOM generation | ❌ | ✅ CycloneDX 1.6 | | Shadow AI detection | ❌ | ✅ Approved model registry | | Evidence export | ❌ | ✅ Signed bundles for auditors | | Incident replay | ❌ | ✅ Full reconstruction + chain verify |
Architecture

flowchart TD
A[Agents / SDKs / Apps\nAI Agents · LangChain · SaaS · Internal Tools] --> B
subgraph B[AIR Blackbox Gateway]
B1[Policy Engine]
B2[Prompt Vault]
B3[Security Scanner]
B4[Compliance Checker]
end
B --> C
subgraph C[Execution Layer]
C1[LLM Providers\nOpenAI · Anthropic · Local Models]
C2[Agent Tools]
C3[External APIs]
end
C --> D
subgraph D[Observability Fabric]
D1[Trace Regression]
D2[Agent Episode Store]
D3[Runtime AI-BOM]
D4[OpenTelemetry]
end
D --> E[Developer Tools\nDebug · Audit · Replay]
How It Works
Think of AIR Blackbox as the control tower for AI agents. Instead of agents calling LLMs and tools directly, everything flows through the gateway.
1 — Entry Layer Agents, SDKs, or applications send requests through the gateway. Supports AI agents, LangChain apps, internal tools, and SaaS platforms.
2 — Safety + Governance Layer Before anything executes, the gateway checks:
| Module | Purpose | |---|---| | Policy Engine | Rules for agent behavior | | Prompt Vault | Versioned prompt templates | | Security Scanner | Prevents prompt injection | | Compliance Checker | EU AI Act regulatory signals |
3 — Execution Layer Once validated, the gateway orchestrates LLM calls, tools, external APIs, and data access.
4 — Observability Fabric Every action is recorded. This is where AIR Blackbox shines:
- Trace regression harness
- Agent episode store
- Runtime AI-BOM emitter
- OpenTelemetry traces
Enables debugging, replaying failures, compliance logs, and model governance.
5 — Developer Interface Replay agent decisions, inspect prompt chains, audit policy decisions, and analyze traces.
Agents / Apps
│
▼
AIR Blackbox Gateway
│
┌─────────────┐
│ Governance │
│ Security │
│ Compliance │
└─────────────┘
│
▼
Agent Runtime (LLMs + Tools)
│
▼
Observability + Replay
File Structure
gateway/
├── cmd/ # Go proxy binary + replayctl CLI
├── collector/ # Go gateway core
├── pkg/ # Go shared packages (trust, audit chain)
├── sdk/ # Python SDK (
