Ember
Production-ready AI agent framework enabling organizations to deploy autonomous agents that are powerful, extensible, and inherently safe. Built on the GAME methodology, Ember provides the runtime foundation for agents that operate within defined boundaries while maintaining the flexibility to solve complex, multi-step problems.

Built for Production AI Agents
Ember implements the GAME methodology, providing a complete foundation for building intelligent, autonomous agents.
GAME Architecture
Built on the Goals, Actions, Memories, Environment methodology. Define weighted goals with reflection, execute plans via tools and LLM, maintain multi-tier memory with semantic search, and enforce environmental constraints declaratively.
Bounded Autonomy
Agents operate autonomously within explicitly defined boundaries. Every agent declares its constraints in YAML—max reasoning depth, tool allowlists, HTTP method gating, and prompt injection defense—all enforced at the framework level, not as an afterthought.
Multi-Tier Memory System
Ephemeral scratchpad, sliding window session memory, and long-term vector storage for semantic search. Shared memory pool with private, workspace, task, and global scopes—tenant-isolated, policy-checked, versioned, and searchable.
Multi-Agent Coordination
Deploy fleets of specialized agents with master/worker roles. Structured handoff protocol with capability discovery, load balancing, and context transfer. Causal trace propagation flows through handoffs, jobs, and memory ops for full interaction graphs.
Native Tool Calling & Plugins
Rich metadata with auto-generated JSON schemas from type hints. Built-in HTTP client with domain allowlists, MCP client support, and extensible @tool decorator. Drop-in plugins with automatic discovery and type-safe validation.
Deterministic Failure Handling
Every failure path is explicit, bounded, and recoverable. Auto-retry with exponential backoff, dead letter queues for forensic analysis, HTTP transport fallback to shared memory, and configurable fail-mode (open/closed). No silent failures, no undefined behavior.
Governance-Aware From Boot
When Lens is enabled, every agent operation is evaluated against active policies in real-time. Sub-millisecond overhead means governance never slows agents down.
Sub-ms Overhead
Lens SDK in-process evaluation adds <1ms to operations
Fail-Mode Config
Fail-open for resilience, fail-closed for high-security
Tamper-Evident Audit
Merkle chain records every policy decision
No Agent Override
Agents cannot bypass or modify policy verdicts
Persistent Visibility
Ember agents report their state continuously to both Lens and FireDeck, providing operators with real-time awareness of their entire agent fleet.
Every 30 seconds, agents report liveness, risk score, and request count
On startup, agents register with model ID, cluster identity, and capabilities
Per-agent and per-cluster risk scores updated with every heartbeat
Visualize handoff patterns between agents in real-time
Automatic detection of persona drift, goal drift, and performance degradation
Quarantine, terminate, and HITL escalation paths for problematic agents
Multi-Protocol API Access
Expose your agents via HTTP with JSON-RPC 2.0, REST API, and WebSocket support. Built-in authentication, rate limiting, and webhook delivery.
JSON-RPC 2.0
Standard protocol support
REST API
RESTful endpoints
WebSocket
Real-time streaming
Admin API
Management endpoints
Flexible Model Support
Multi-provider support: OpenAI, Anthropic, OpenRouter.
OpenAI
GPT-5, GPT-4, GPT-4o, GPT-3.5-turbo
Anthropic
Claude 4.x, Claude 3.x, Sonnet, Opus
OpenRouter
100+ models via unified API
Production-Ready Deployment
Deploy Ember agents anywhere with flexible options for containers, orchestration, and microVMs.
Single Agent
One daemon, one agent via Docker Compose—simplest path to production
Multi-Agent Cluster
Generate fleet deployments from a single YAML with master/worker roles and shared infrastructure
Firecracker VM
Lightweight microVM deployment for maximum isolation on Linux hosts with KVM
Scalable Fleets
Individual agents have a minimal memory footprint, permitting dozens on a typical host
Cluster-Level Safety Guarantees
Multi-agent cluster mode enables deploying fleets of specialized agents that coordinate safely with role-based authority and containment guarantees.
Master agents orchestrate; worker agents execute within boundaries. Roles enforced at framework level.
Each agent in its own Docker container with CPU, memory, PID limits, and log rotation.
Each agent dequeues only its own jobs—no cross-agent interference.
Private, workspace, task, and global scopes with tenant-isolated access controls.
Failed jobs quarantined for forensic analysis, never silently dropped.
Queue depth limits and payload size limits prevent resource exhaustion.
Comprehensive Audit Logging
Structured JSON logging with 12+ event types, automatic sensitive data redaction, request correlation, and performance metrics for complete visibility.
Built for Every Team
AI Engineers
- Drop-in framework—YAML config + Python API
- Plugin system—add tools without touching core
- Multi-provider: OpenAI, Anthropic, OpenRouter
- Type-safe validation for all configuration
Platform Teams
- Cluster deployment—generate Docker Compose from a single YAML
- Resource governance—CPU, memory, PID limits per agent
- Observability—structured JSON audit logs, heartbeats, risk scores
- Self-healing—auto-restart, auto-retry, dead letter queues
Security & Compliance
- Bounded autonomy—agents cannot exceed declared boundaries
- Audit trail—every operation logged with session/request correlation
- Policy enforcement—Lens integration for real-time governance
- Containment—quarantine, terminate, and HITL escalation paths
Why Ember?
| Capability | Generic Frameworks | Ember |
|---|---|---|
| Safety Model | None (DIY) | Built-in bounded autonomy + Lens governance |
| Multi-Agent Coordination | Basic message passing | Structured handoff protocol with role enforcement |
| Memory Sharing | None or ad-hoc | Tenant-isolated shared pool with scopes and versioning |
| Resource Limits | None | Container-level CPU/memory/PID limits |
| Failure Handling | Crash and hope | Deterministic: retry → dead letter → alert |
| Audit Logging | Basic stdout | Structured JSON with 12+ event types, PII redaction |
| Deployment | Manual | One-command cluster generation from YAML |
| Governance | Bolt-on | Native Lens SDK, policy-aware from boot |
Native Lens Integration
Ember is governance-aware from boot. When Lens is enabled, every tool call, memory access, and handoff is evaluated against active policies in real-time with sub-millisecond overhead and tamper-evident audit logging.
Native FireDeck Integration
Ember agents authenticate, register, and heartbeat with FireDeck-issued credentials. Bind agents to verifiable identities, manage credential lifecycles, track per-agent risk posture, and orchestrate workloads across your entire agent fleet.