A RAG pipeline that works for a simple chatbot is not enough for multi-agent AI systems. This live workshop shows you how to build high-fidelity RAG with citations, memory engineering, and multi-agent integration — the retrieval layer that makes production AI systems accurate.
By Packt Publishing · Refunds up to 10 days before
Standard RAG retrieves documents and passes them to a single LLM. Multi-agent RAG must handle multiple agents querying the same knowledge base, maintain citation chains across agent handoffs, manage context window constraints per agent, and prevent conflicting retrievals from causing coordination failures. This workshop builds RAG designed for this complexity.
Context engineering is the discipline of designing systems that give AI the right information, in the right format, to reason and act reliably. It goes beyond prompt engineering — building structured, deterministic systems that scale in production rather than depending on fragile prompts.
A multi-agent system uses multiple specialized AI agents working together — each with a defined role, context, and tools — to complete complex tasks no single agent could handle reliably. Context engineering is the key to making them work predictably.
MCP is Anthropic's open standard for connecting AI models to tools, data sources, and other agents. It provides a structured way to orchestrate multi-agent workflows with clear context boundaries — making systems transparent and debuggable.
Context engineering concepts require hands-on practice to truly understand. This live workshop lets you build a working system with a world-class instructor answering your questions in real time — far more effective than reading documentation alone.
Six modules. Six hours. A production-ready context-engineered AI system by the time you finish.
Understand why prompts fail at scale and how semantic blueprints give AI systems structured, goal-driven contextual awareness that scales reliably.
Design and orchestrate multi-agent workflows using the Model Context Protocol. Build transparent, traceable agent systems that coordinate reliably.
Build retrieval augmented generation pipelines that deliver accurate, cited responses. Engineer memory systems that persist context reliably across agent interactions.
Architect a transparent, explainable context engine where every decision is traceable. Build AI systems that are predictable and debuggable in production.
Implement safeguards against prompt injection and data poisoning. Enforce moderation, trust boundaries, and access controls in multi-agent environments.
Deploy your context-engineered multi-agent system to production. Apply patterns for scaling, monitoring, and maintaining reliability under real-world load.
Concrete working deliverables — not just theory and slides.
A working Glass-Box Context Engine with transparent, traceable reasoning
Multi-agent workflow orchestrated with the Model Context Protocol
High-fidelity RAG pipeline with memory and citations
Safeguards against prompt injection and data poisoning
Reusable architecture patterns for production AI systems
Certificate of completion from Packt Publishing
Denis Rothman brings decades of production AI engineering experience to this live workshop — making complex context engineering concepts immediately actionable.
Denis Rothman is a bestselling AI author with over 30 years of experience in artificial intelligence, agent systems, and optimization. He has authored multiple cutting-edge AI books published by Packt and is renowned for making complex AI architecture concepts practical and immediately applicable. In this workshop he guides you step by step through building production-ready context-engineered multi-agent systems — answering your questions live throughout the 6-hour session.
This is an intermediate to advanced workshop. Solid Python and basic LLM experience required.
Common questions about the workshop, what to expect, and how to prepare.
A RAG pipeline designed for multi-agent AI needs several properties beyond basic retrieval: citation tracking that persists through agent handoffs, per-agent context management that respects individual agent context windows, shared memory that multiple agents can read and write consistently, conflict resolution for contradictory retrievals, and integration with MCP for typed retrieval requests and responses. This workshop builds all of these.
The workshop covers citation tracking at every layer of the RAG pipeline — storing source metadata with embeddings, attaching source attribution to retrieved chunks, propagating citations through the LLM response generation, and verifying that claims in the final output trace back to specific retrieved sources. This citation chain makes the entire system auditable.
Memory engineering in a multi-agent RAG pipeline involves three layers: working memory (the current agent context window), episodic memory (conversation history across sessions), and semantic memory (the embedded knowledge base). The workshop covers how to manage all three layers so agents have appropriate access to past interactions, structured knowledge, and current task context without context overflow.
Chatbot RAG is a stateless retrieval at each turn. Agent RAG must maintain state across agent handoffs, coordinate access to the knowledge base across multiple simultaneous agents, track which agent retrieved which information, and manage the context overhead that multiple retrieval operations add to each agent's context window. The workshop builds agent-grade RAG from the start.
RAG hallucination prevention in this workshop operates at two levels: retrieval quality (using confidence scoring and source verification to surface only reliable documents) and generation quality (using citation-grounded prompting that requires the model to attribute every claim to a specific retrieved source). Uncited claims trigger a validation step before they propagate through the multi-agent system.
Yes. The RAG architecture taught in this workshop is designed to be vector database agnostic. The retrieval interface is abstracted through MCP, so you can swap the underlying vector store without changing the agent orchestration layer. The instructor discusses vector database selection considerations for production RAG systems during the workshop.
6 hours. Bestselling AI author. Production context-engineered multi-agent system by the end. Seats are limited.
Register Now →Saturday April 25 · 9am to 3pm EDT · Online · Packt Publishing · Cohort 2