Context window management is the difference between AI agents that stay sharp across long interactions and those that degrade after a few turns. This live workshop teaches the memory engineering techniques that keep every agent's context window filled with exactly the right information at every step.
By Packt Publishing · Refunds up to 10 days before
Every LLM agent has a finite context window. In a multi-agent system, that window must be shared between the semantic blueprint, retrieved knowledge, conversation history, and inter-agent results. Without explicit management, the window fills with stale and irrelevant content, causing the agent to lose coherence. This workshop builds the management layer that prevents this.
Context engineering is the discipline of designing systems that give AI the right information, in the right format, to reason and act reliably. It goes beyond prompt engineering — building structured, deterministic systems that scale in production.
A multi-agent system uses multiple specialised AI agents working together — each with a defined role, context, and tools — to complete complex tasks no single agent could handle reliably. Context engineering makes them predictable.
MCP is Anthropic's open standard for connecting AI models to tools, data sources, and other agents. It provides structured agent orchestration with clear context boundaries — making systems transparent and debuggable.
Context engineering requires hands-on practice to truly understand. This live workshop lets you build a working system with a world-class instructor answering your questions in real time.
Six modules. Six hours. A production-ready context-engineered AI system by the time you finish.
Understand why prompts fail at scale and how semantic blueprints give AI structured, goal-driven contextual awareness.
Design and orchestrate multi-agent workflows using the Model Context Protocol. Build transparent, traceable agent systems.
Build RAG pipelines that deliver accurate, cited responses. Engineer memory systems that persist context reliably across agents.
Architect a transparent, explainable context engine where every decision is traceable and debuggable in production.
Implement safeguards against prompt injection and data poisoning. Enforce trust boundaries in multi-agent environments.
Deploy your context-engineered system to production. Apply patterns for scaling, monitoring, and reliability.
Concrete working deliverables — not just theory and slides.
A working Glass-Box Context Engine with transparent, traceable reasoning
Multi-agent workflow orchestrated with the Model Context Protocol
High-fidelity RAG pipeline with memory and citations
Safeguards against prompt injection and data poisoning
Reusable architecture patterns for production AI systems
Certificate of completion from Packt Publishing
Denis Rothman brings decades of production AI engineering experience to this live workshop.
Denis Rothman is a bestselling AI author with over 30 years of experience in artificial intelligence, agent systems, and optimization. He has authored multiple cutting-edge AI books published by Packt and is renowned for making complex AI architecture concepts practical and immediately applicable. He guides you step by step through building production-ready context-engineered multi-agent systems — answering your questions live throughout the 6-hour session.
Intermediate to advanced workshop. Solid Python and basic LLM experience required.
Everything you need to know before registering.
Context window management is the systematic control of what content occupies each agent's context window at every invocation. In a multi-agent system, poor context window management causes: agents receiving irrelevant content from earlier in a conversation (context pollution), agents losing track of important information as the window fills (context overflow), and agents in different parts of the workflow working from different context states (context inconsistency). Explicit management prevents all three problems.
A context budget allocates the available context window tokens among different content types: a fixed allocation for the semantic blueprint, an allocation for RAG retrieved content, an allocation for conversation history, an allocation for inter-agent results, and a reserve for the agent's response. The context manager enforces these budgets by truncating or compressing content that exceeds its allocation before the context is assembled for agent invocation. The workshop covers implementing a token-counting context budget manager in Python.
The most effective context compression techniques for multi-agent AI are: conversation summarisation (replacing detailed conversation history with a compact semantic summary that preserves key decisions and facts), document chunking with relevance scoring (retrieving only the most relevant sections of long documents rather than full content), and episodic memory encoding (compressing past agent interactions into structured records that can be retrieved selectively rather than replayed in full). The workshop implements all three as production-ready Python components.
The semantic blueprint specifies the categories of context each agent needs. The context window manager fills those categories from available sources (RAG retrieval, episodic memory, conversation history) while respecting the budget allocated for each. Relevance scoring determines which specific content items are included when the available content exceeds the budget allocation. The workshop covers the relevance scoring and content selection logic that makes context window allocation decisions systematic.
Large context windows reduce overflow frequency but introduce a different problem: models lose focus on the most important information when it is buried deep in a large context. Context window management for large-window models focuses on positioning: placing the most important content (current task specification, highest-relevance retrievals) in the positions where the model attends most reliably, and using structured content markers that help the model navigate the large context effectively. The workshop covers position-aware context management.
Context window efficiency is measured through the Glass-Box logging layer by tracking: token utilization per category (blueprint, RAG, history, inter-agent results), content relevance scores for retrieved items that entered the context versus those that were filtered out, citation utilization (what percentage of retrieved content was actually referenced in the agent's output), and context overflow events (invocations where the available content exceeded the total budget before selection). These metrics reveal where context window management can be improved.
6 hours. Bestselling AI author. Production context-engineered multi-agent system by the end. Seats are limited.
Register Now →Saturday April 25 · 9am to 3pm EDT · Online · Packt Publishing · Cohort 2