Production deployment workflow for agentic systems. Directs to RAG for implementation.
| Framework | Primary Option | Alternative | RAG Query |
|---|---|---|---|
| ADK | Agent Engine (Vertex AI) | Cloud Run, GKE | "ADK deployment agent engine" |
| OpenAI | Any Python hosting | Serverless, Docker | "openai agents deployment" |
| LangChain | LangServe, Cloud Run | Docker, K8s | "langchain langserve deployment" |
| LangGraph | LangGraph Platform | Cloud Run | "langgraph platform deployment" |
| CrewAI | CrewAI Enterprise | Docker | "crewai deployment production" |
| Anthropic | Any Python hosting | Docker, Serverless | "anthropic agent deployment" |
RAG Query: mcp__agentic-rag__search("[framework] environment configuration", mode="explain")
Production differs from development:
RAG Query: mcp__agentic-rag__search("[framework] dockerfile", mode="build")
RAG Query: mcp__agentic-rag__search("[framework] [platform] deployment", mode="explain")
RAG Query: mcp__agentic-rag__search("[framework] monitoring observability", mode="explain")
| Metric | Alert Threshold | Why It Matters |
|---|---|---|
| Latency p95 | > 5s | User experience |
| Error rate | > 1% | Reliability |
| Token usage | Spike > 200% | Cost control |
| Tool failures | > 5% | Agent effectiveness |
| Routing accuracy | < 90% | Multi-agent health |
RAG Query: mcp__agentic-rag__search("agent input validation security", mode="explain")
RAG Query: mcp__agentic-rag__search("agent guardrails output filtering", mode="explain")
RAG Query: mcp__agentic-rag__search("[framework] secret management", mode="explain")
| Concern | Solution | RAG Query |
|---|---|---|
| Cold starts | Keep warm instances | "[framework] cold start" |
| Concurrent requests | Queue + workers | "[framework] scaling" |
| Token limits | Request batching | "[framework] rate limiting" |
| State persistence | External store | "[framework] state persistence" |