Docker for AI Agents
An open spec, runtime, and toolchain for packaging AI agents as portable, observable, deployable artifacts.
Define agents in YAML. Bundle knowledge bases with vector embeddings. Deploy to any infrastructure. Observe every request.
Define your agent
Write a YAML spec with your system prompt, skills, knowledge base reference, and merge policies. It's your agent's portable definition.
apiVersion: arachne.ai/v0
kind: Agent
metadata:
name: support-agent
spec:
model: gpt-4o
systemPrompt: |
You are a helpful support agent.
Answer using the knowledge base.
knowledgeBaseRef: product-docs Build and push
Weave your agent into a deployable artifact with embedded knowledge base. Push to any Arachne registry.
$ arachne weave agent.yaml Weaving support-agent... Chunking 42 documents (650 tokens, 120 overlap) Embedding 186 chunks via text-embedding-3-small Packaged support-agent.orb (2.4 MB) $ arachne push support-agent.orb Pushed myorg/support-agent:latest
Deploy and observe
Deploy to any environment. Every request is traced with latency, token usage, and RAG retrieval metrics. Zero additional code required.
$ arachne deploy myorg/support-agent:latest --env production
Deployed to production
$ curl https://api.arachne-ai.com/v1/chat/completions \
-H "Authorization: Bearer loom_sk_..." \
-d '{"model":"gpt-4o","messages":[...]}' Why Arachne?
Without Arachne
- ✗ Glue code for every LLM provider
- ✗ Separate vector DB setup and management
- ✗ Custom observability pipeline
- ✗ Conversation state managed in app code
- ✗ Agent configs scattered across services
- ✗ No standard way to package or deploy
With Arachne
- ✓ One API for OpenAI, Azure, Ollama
- ✓ Embedded vector search with pgvector
- ✓ Tracing built into the gateway
- ✓ Conversation memory as a runtime feature
- ✓ Portable YAML agent specs
- ✓
arachne weave → push → deploy
What the Runtime Gives You
Everything you need to ship AI agents
Conversation Memory
Automatic conversation history with token-aware summarization. Enable per-agent with one flag. History persists across requests with partition support.
RAG Knowledge Bases
Upload documents, auto-chunk and embed them, then attach to any agent. Vector search retrieves relevant context for every request automatically.
Full Observability
Every request traced with latency, token usage, model, and provider. Fire-and-forget recording with zero gateway overhead. All data encrypted at rest.
Multi-Provider Routing
OpenAI, Azure OpenAI, and Ollama through a single API. Switch providers without changing client code. Per-agent or per-tenant configuration.
Portable Agent Spec
Define agents in YAML with system prompts, skills, merge policies, and knowledge base refs. Weave into artifacts. Push to any registry.
Multi-Tenant by Default
Built-in tenant isolation with subtenant hierarchies, per-tenant encryption, role-based access, and invite-based onboarding.
Join the Beta
Get early access to Arachne. We'll send you an invite when your account is ready.