AI agent memory is one of the hardest engineering problems in production AI. This live workshop teaches the three-layer memory architecture that makes agents retain useful context without context window overflow.
By Packt Publishing · Refunds up to 10 days before
Most AI agent memory approaches are either too simple (only current context) or too complex (storing everything). The memory engineering approach taught in this workshop designs a three-layer system that gives agents precisely the memory they need for reliable, long-running interactions.
Context engineering is the discipline of designing systems that give AI the right information, in the right format, to reason and act reliably. It goes beyond prompt engineering — building structured, deterministic systems that scale in production.
A multi-agent system uses multiple specialised AI agents working together — each with a defined role, context, and tools — to complete complex tasks no single agent could handle reliably. Context engineering makes them predictable.
MCP is Anthropic's open standard for connecting AI models to tools, data sources, and other agents. It provides structured agent orchestration with clear context boundaries — making systems transparent and debuggable.
Context engineering requires hands-on practice to truly understand. This live workshop lets you build a working system with a world-class instructor answering your questions in real time.
Six modules. Six hours. A production-ready context-engineered AI system by the time you finish.
Understand why prompts fail at scale and how semantic blueprints give AI structured, goal-driven contextual awareness.
Design and orchestrate multi-agent workflows using the Model Context Protocol. Build transparent, traceable agent systems.
Build RAG pipelines that deliver accurate, cited responses. Engineer memory systems that persist context reliably across agents.
Architect a transparent, explainable context engine where every decision is traceable and debuggable in production.
Implement safeguards against prompt injection and data poisoning. Enforce trust boundaries in multi-agent environments.
Deploy your context-engineered system to production. Apply patterns for scaling, monitoring, and reliability.
Concrete working deliverables — not just theory and slides.
A working Glass-Box Context Engine with transparent, traceable reasoning
Multi-agent workflow orchestrated with the Model Context Protocol
High-fidelity RAG pipeline with memory and citations
Safeguards against prompt injection and data poisoning
Reusable architecture patterns for production AI systems
Certificate of completion from Packt Publishing
Denis Rothman brings decades of production AI engineering experience to this live workshop.
Denis Rothman is a bestselling AI author with over 30 years of experience in artificial intelligence, agent systems, and optimization. He has authored multiple cutting-edge AI books published by Packt and is renowned for making complex AI architecture concepts practical and immediately applicable. He guides you step by step through building production-ready context-engineered multi-agent systems — answering your questions live throughout the 6-hour session.
This is an intermediate to advanced workshop. Solid Python and basic LLM experience required.
Everything you need to know before registering.
This workshop covers three memory layers: working memory (the active context window for the current task), episodic memory (a compressed record of past interactions that can be selectively retrieved), and semantic memory (the embedded knowledge base accessed through the RAG pipeline). Designing these three layers to work together is what gives AI agents reliable long-term context without context window overflow.
Context window overflow is prevented through active memory management: summarising and compressing episodic memory rather than retaining raw transcripts, using selective retrieval to pull only relevant memories into working memory, and setting explicit context budgets per agent that trigger compression when approaching the threshold. The workshop covers all three techniques with practical Python implementation.
Short-term AI agent memory is working memory: the current context window contents available for immediate use. Long-term memory is episodic and semantic memory: past interactions stored in compressed form and a knowledge base, both accessible through retrieval. The engineering challenge is moving information efficiently between these layers without losing important context or overwhelming working memory.
RAG serves as the retrieval interface to semantic memory. When an agent needs domain knowledge or past context not in its current working memory, it queries the RAG pipeline which retrieves relevant content from the embedded knowledge base. The retrieved content is injected into working memory as structured citation, keeping context relevant and verifiable.
Yes. The episodic memory layer is designed to persist across sessions, storing compressed conversation summaries and key decisions in a retrievable format. When a new session begins, the memory system retrieves relevant episodic memories to give the agent appropriate context from past interactions. The workshop covers session persistence implementation for production systems.
Memory sharing between agents requires explicit access controls and clear versioning to prevent one agent's writes from corrupting another's state. The workshop covers shared memory architecture using the MCP resource system, which provides typed, validated read and write access to shared memory stores. Each agent's access is logged by the Glass-Box layer making memory interactions auditable.
6 hours. Bestselling AI author. Production context-engineered multi-agent system by the end. Seats are limited.
Register Now →Saturday April 25 · 9am to 3pm EDT · Online · Packt Publishing · Cohort 2