RAG pipelines designed for chatbots do not work well in multi-agent systems. Concurrent agent access, citation propagation across agent handoffs, and shared memory management require a different RAG architecture. This live workshop builds RAG engineered specifically for multi-agent use.
By Packt Publishing · Refunds up to 10 days before
Chatbot RAG is stateless retrieval per turn. Multi-agent RAG must handle multiple agents querying the same knowledge base simultaneously, propagate citation chains through agent handoffs, manage shared memory across agents, and expose retrieval as a composable MCP service. This workshop builds each of these capabilities.
Context engineering is the discipline of designing systems that give AI the right information, in the right format, to reason and act reliably. It goes beyond prompt engineering — building structured, deterministic systems that scale in production.
A multi-agent system uses multiple specialised AI agents working together — each with a defined role, context, and tools — to complete complex tasks no single agent could handle reliably. Context engineering makes them predictable.
MCP is Anthropic's open standard for connecting AI models to tools, data sources, and other agents. It provides structured agent orchestration with clear context boundaries — making systems transparent and debuggable.
Context engineering requires hands-on practice to truly understand. This live workshop lets you build a working system with a world-class instructor answering your questions in real time.
Six modules. Six hours. A production-ready context-engineered AI system by the time you finish.
Understand why prompts fail at scale and how semantic blueprints give AI structured, goal-driven contextual awareness.
Design and orchestrate multi-agent workflows using the Model Context Protocol. Build transparent, traceable agent systems.
Build RAG pipelines that deliver accurate, cited responses. Engineer memory systems that persist context reliably across agents.
Architect a transparent, explainable context engine where every decision is traceable and debuggable in production.
Implement safeguards against prompt injection and data poisoning. Enforce trust boundaries in multi-agent environments.
Deploy your context-engineered system to production. Apply patterns for scaling, monitoring, and reliability.
Concrete working deliverables — not just theory and slides.
A working Glass-Box Context Engine with transparent, traceable reasoning
Multi-agent workflow orchestrated with the Model Context Protocol
High-fidelity RAG pipeline with memory and citations
Safeguards against prompt injection and data poisoning
Reusable architecture patterns for production AI systems
Certificate of completion from Packt Publishing
Denis Rothman brings decades of production AI engineering experience to this live workshop.
Denis Rothman is a bestselling AI author with over 30 years of experience in artificial intelligence, agent systems, and optimization. He has authored multiple cutting-edge AI books published by Packt and is renowned for making complex AI architecture concepts practical and immediately applicable. He guides you step by step through building production-ready context-engineered multi-agent systems — answering your questions live throughout the 6-hour session.
This is an intermediate to advanced workshop. Solid Python and basic LLM experience required.
Everything you need to know before registering.
A multi-agent RAG pipeline differs from chatbot RAG in four key ways: it handles concurrent agent queries with connection pooling and caching, it tracks citation chains through agent handoffs so downstream agents know the original source of any claim, it provides a shared episodic memory store that multiple agents can read and write, and it exposes all retrieval capabilities as MCP-typed services that agents can invoke through the standard protocol.
When agent A retrieves a document and uses it in its output, that citation is attached to the output as structured metadata. When agent B processes agent A's output, the citation metadata propagates with the claimed fact. This citation chain means the final output of the multi-agent system can trace every factual claim back to the original retrieved source, regardless of how many agents processed it along the way.
Shared RAG access in a multi-agent system uses connection pooling to handle concurrent queries, a read-through cache to avoid redundant embedding searches for similar queries, and distributed locking for any write operations to the knowledge base. The workshop covers the connection management and caching architecture that makes shared RAG access reliable and performant under concurrent agent load.
Exposing RAG as an MCP service means defining an MCP server with retrieval tools (query the knowledge base, retrieve by ID), resource endpoints (access specific documents), and prompt templates (structure retrieval results for agent consumption). The workshop covers the complete MCP RAG server implementation that lets any agent in the system invoke retrieval through the standard protocol.
The workshop covers three caching strategies for multi-agent RAG: query result caching (storing retrieval results for identical queries), embedding caching (pre-computing embeddings for frequently accessed documents), and semantic caching (storing results for semantically similar queries even if the exact text differs). Each caching layer reduces latency and improves consistency for concurrent agent access.
Keeping the knowledge base current while agents are running requires an incremental indexing strategy that adds new content to the vector store without requiring a full re-index. The workshop covers implementing a document change monitoring pipeline that detects updated content, re-embeds only changed documents, and updates the index while maintaining read availability for agents that are actively querying.
6 hours. Bestselling AI author. Production context-engineered multi-agent system by the end. Seats are limited.
Register Now →Saturday April 25 · 9am to 3pm EDT · Online · Packt Publishing · Cohort 2