Running MCP in development is straightforward. Running it in production requires containerisation, service discovery, health monitoring, graceful failure handling, and zero-downtime updates. This live workshop covers the complete production MCP deployment from infrastructure to monitoring.
By Packt Publishing · Refunds up to 10 days before
Production MCP deployment goes far beyond starting a server locally. It requires containerised agent servers with health checks, a service discovery mechanism for the orchestrator, Glass-Box monitoring for every tool invocation, graceful degradation when agents fail, and deployment pipelines that update agents without disrupting running workflows. This workshop covers all of it.
Context engineering is the discipline of designing systems that give AI the right information, in the right format, to reason and act reliably. It goes beyond prompt engineering — building structured, deterministic systems that scale in production.
A multi-agent system uses multiple specialised AI agents working together — each with a defined role, context, and tools — to complete complex tasks no single agent could handle reliably. Context engineering makes them predictable.
MCP is Anthropic's open standard for connecting AI models to tools, data sources, and other agents. It provides structured agent orchestration with clear context boundaries — making systems transparent and debuggable.
Context engineering requires hands-on practice to truly understand. This live workshop lets you build a working system with a world-class instructor answering your questions in real time.
Six modules. Six hours. A production-ready context-engineered AI system by the time you finish.
Understand why prompts fail at scale and how semantic blueprints give AI structured, goal-driven contextual awareness.
Design and orchestrate multi-agent workflows using the Model Context Protocol. Build transparent, traceable agent systems.
Build RAG pipelines that deliver accurate, cited responses. Engineer memory systems that persist context reliably across agents.
Architect a transparent, explainable context engine where every decision is traceable and debuggable in production.
Implement safeguards against prompt injection and data poisoning. Enforce trust boundaries in multi-agent environments.
Deploy your context-engineered system to production. Apply patterns for scaling, monitoring, and reliability.
Concrete working deliverables — not just theory and slides.
A working Glass-Box Context Engine with transparent, traceable reasoning
Multi-agent workflow orchestrated with the Model Context Protocol
High-fidelity RAG pipeline with memory and citations
Safeguards against prompt injection and data poisoning
Reusable architecture patterns for production AI systems
Certificate of completion from Packt Publishing
Denis Rothman brings decades of production AI engineering experience to this live workshop.
Denis Rothman is a bestselling AI author with over 30 years of experience in artificial intelligence, agent systems, and optimization. He has authored multiple cutting-edge AI books published by Packt and is renowned for making complex AI architecture concepts practical and immediately applicable. He guides you step by step through building production-ready context-engineered multi-agent systems — answering your questions live throughout the 6-hour session.
Intermediate to advanced workshop. Solid Python and basic LLM experience required.
Everything you need to know before registering.
Production MCP infrastructure requires: containerised MCP servers (typically Docker) with health check endpoints that the orchestrator polls to verify availability, a service registry where agent servers register their addresses and capabilities, a load balancer for MCP servers that need to handle high request volume, persistent storage for the Glass-Box logs and episodic memory, and monitoring infrastructure that tracks tool invocation latency and error rates per agent server.
Containerising MCP agent servers uses Docker with a Python base image, the MCP server dependencies installed from a pinned requirements file, health check configuration that verifies the server can accept tool invocations, and resource limit configuration that prevents any single server from consuming excessive compute. The workshop covers a Docker configuration pattern for MCP servers that works across development, staging, and production environments.
Production MCP server health checks verify that the server can accept and respond to tool invocations within an acceptable latency threshold. A simple health check tool that echoes a test payload is registered on each server and polled by the orchestrator's availability monitor. Servers that fail health checks are temporarily removed from the orchestrator's active server list until they recover, preventing the orchestrator from routing tasks to unavailable agents.
Zero-downtime MCP server updates use a blue-green deployment pattern: the new server version is deployed and health-checked alongside the running version, the orchestrator's service registry is updated to route new requests to the updated server, running requests on the old server are allowed to complete, and the old server is shut down once its request queue is empty. This pattern requires the orchestrator to track which server version each active request was dispatched to.
Production MCP monitoring covers: tool invocation latency histograms per server and tool (to detect performance regressions), error rates by error type (to distinguish transient from persistent failures), connection pool utilization (to detect capacity issues), Glass-Box trace completeness (to verify no requests are dropping steps), and semantic drift metrics (to detect when agent behavior changes in ways not reflected in error rates).
Horizontal scaling of MCP agent servers uses multiple instances behind a load balancer for stateless agents, consistent hashing for agents that maintain per-request state, and auto-scaling policies based on request queue depth and tool invocation latency. The workshop covers the stateless agent design pattern that makes horizontal scaling straightforward, and the session affinity patterns needed for the few agent types that require state continuity across tool invocations.
6 hours. Bestselling AI author. Production context-engineered multi-agent system by the end. Seats are limited.
Register Now →Saturday April 25 · 9am to 3pm EDT · Online · Packt Publishing · Cohort 2