A context engine is the core infrastructure that makes AI agents reliable. This live workshop builds a complete context engine from scratch: the semantic blueprint generator, context router, memory manager, RAG integration, and the Glass-Box observability layer, all connected into a working production system.
By Packt Publishing · Refunds up to 10 days before
You will build a complete Glass-Box Context Engine in Python: five interconnected components (blueprint generator, context router, MCP orchestrator, RAG pipeline, memory manager) connected by a Glass-Box logging layer that makes every decision observable. By the end of the 6-hour session, the engine is running and orchestrating multi-agent workflows.
Context engineering is the discipline of designing systems that give AI the right information, in the right format, to reason and act reliably. It goes beyond prompt engineering — building structured, deterministic systems that scale in production.
A multi-agent system uses multiple specialised AI agents working together — each with a defined role, context, and tools — to complete complex tasks no single agent could handle reliably. Context engineering makes them predictable.
MCP is Anthropic's open standard for connecting AI models to tools, data sources, and other agents. It provides structured agent orchestration with clear context boundaries — making systems transparent and debuggable.
Context engineering requires hands-on practice to truly understand. This live workshop lets you build a working system with a world-class instructor answering your questions in real time.
Six modules. Six hours. A production-ready context-engineered AI system by the time you finish.
Understand why prompts fail at scale and how semantic blueprints give AI structured, goal-driven contextual awareness.
Design and orchestrate multi-agent workflows using the Model Context Protocol. Build transparent, traceable agent systems.
Build RAG pipelines that deliver accurate, cited responses. Engineer memory systems that persist context reliably across agents.
Architect a transparent, explainable context engine where every decision is traceable and debuggable in production.
Implement safeguards against prompt injection and data poisoning. Enforce trust boundaries in multi-agent environments.
Deploy your context-engineered system to production. Apply patterns for scaling, monitoring, and reliability.
Concrete working deliverables — not just theory and slides.
A working Glass-Box Context Engine with transparent, traceable reasoning
Multi-agent workflow orchestrated with the Model Context Protocol
High-fidelity RAG pipeline with memory and citations
Safeguards against prompt injection and data poisoning
Reusable architecture patterns for production AI systems
Certificate of completion from Packt Publishing
Denis Rothman brings decades of production AI engineering experience to this live workshop.
Denis Rothman is a bestselling AI author with over 30 years of experience in artificial intelligence, agent systems, and optimization. He has authored multiple cutting-edge AI books published by Packt and is renowned for making complex AI architecture concepts practical and immediately applicable. He guides you step by step through building production-ready context-engineered multi-agent systems — answering your questions live throughout the 6-hour session.
Intermediate to advanced workshop. Solid Python and basic LLM experience required.
Everything you need to know before registering.
The five core components are: (1) the Semantic Blueprint Generator that creates structured agent specifications from task descriptions, (2) the Context Router that assembles the appropriate context package for each agent invocation, (3) the MCP Orchestrator that dispatches tasks to specialised agent servers and collects typed results, (4) the Knowledge Layer comprising the RAG pipeline and memory manager, and (5) the Glass-Box Logger that records every component operation with structured metadata for observability and debugging.
The context engine is built incrementally across the 6-hour session: the Glass-Box Logger is built first (30 minutes) as the foundation for observing everything else, the Semantic Blueprint Generator next (45 minutes), followed by the Knowledge Layer (60 minutes), the MCP Orchestrator (60 minutes), the Context Router (45 minutes), and integration testing and production preparation (the final 60 minutes). Each component is tested before the next is built.
The components connect through a pipeline architecture: a user request enters the MCP Orchestrator which invokes the Semantic Blueprint Generator to create agent specifications, passes those specifications to the Context Router which assembles context packages from the Knowledge Layer, dispatches agent invocations through MCP with the assembled context packages, and collects results. The Glass-Box Logger wraps every interaction between components, creating a complete audit trail of the entire pipeline execution.
When a specialised agent is unavailable (MCP server health check failure or timeout), the context engine's fault tolerance logic activates: it checks whether a fallback agent server is configured for the same capability, routes the task to the fallback if available, records the primary failure and fallback routing in the Glass-Box log, and returns a partial result to the orchestrator if no fallback is available rather than failing the entire workflow. The workshop covers implementing this fault tolerance logic.
Extending the context engine with a new specialised agent involves: implementing the new agent as an MCP server with typed tools, registering the server address in the context engine's agent registry, adding a capability description to the orchestrator's planner blueprint so the orchestrator knows when to route tasks to the new agent, and adding the appropriate knowledge domain to the RAG pipeline if the new agent requires domain-specific retrieval. No changes to the core context engine components are required.
The context engine built in this workshop is architecturally simpler and more transparent than commercial orchestration platforms, but provides the same core capabilities for most use cases: multi-agent task decomposition, typed agent coordination, RAG-grounded knowledge retrieval, memory management, and comprehensive observability. Building it yourself gives you complete understanding of the system, full customisability for your specific use case, and no vendor lock-in or usage fees.
6 hours. Bestselling AI author. Production context-engineered multi-agent system by the end. Seats are limited.
Register Now →Saturday April 25 · 9am to 3pm EDT · Online · Packt Publishing · Cohort 2