Transparent Explainable AI Agent · Live · April 25

Build a Transparent and Explainable AI Agent — The Glass-Box Approach

A transparent AI agent is not just one that logs outputs. It is one where every reasoning step, knowledge retrieval, and coordination decision is observable and explainable. This live workshop builds a truly transparent AI agent using the Glass-Box Context Engine architecture.

Saturday, April 25  9am – 3pm EDT
6 Hours  Hands-on coding
Cohort 2  Intermediate to Advanced

Workshop Details

📅
Date & Time
Saturday, April 25, 2026
9:00am – 3:00pm EDT
Duration
6 Hours · Hands-on
💻
Format
Live Online · Interactive
📚
Level
Intermediate to Advanced
🎓
Includes
Certificate of Completion
Register on Eventbrite →

By Packt Publishing · Refunds up to 10 days before

✦ By Packt Publishing
6 Hours Live Hands-On
Cohort 2 — April 25, 2026
Intermediate to Advanced
Certificate of Completion
Why Trust Packt

Over 20 Years of Helping Developers Build Real Skills

7,500+
Books and video courses published
108
Live workshops hosted on Eventbrite
30+
Years of AI experience — Denis Rothman
100%
Hands-on — real code every session
About This Workshop

What Transparent and Explainable AI Agents Look Like in Practice

Transparent AI agents are increasingly required by enterprise and regulatory environments. The Glass-Box architecture taught in this workshop provides genuine transparency: every semantic blueprint, every retrieved citation, every agent coordination step is logged in a queryable format that supports both real-time debugging and historical auditing.

🧠

What is Context Engineering?

Context engineering is the discipline of designing systems that give AI the right information, in the right format, to reason and act reliably. It goes beyond prompt engineering — building structured, deterministic systems that scale in production.

🤖

What is a Multi-Agent System?

A multi-agent system uses multiple specialised AI agents working together — each with a defined role, context, and tools — to complete complex tasks no single agent could handle reliably. Context engineering makes them predictable.

🔗

What is the Model Context Protocol?

MCP is Anthropic's open standard for connecting AI models to tools, data sources, and other agents. It provides structured agent orchestration with clear context boundaries — making systems transparent and debuggable.

🎯

Why Attend as a Live Workshop?

Context engineering requires hands-on practice to truly understand. This live workshop lets you build a working system with a world-class instructor answering your questions in real time.

Workshop Curriculum

What This 6-Hour Workshop Covers

Six modules. Six hours. A production-ready context-engineered AI system by the time you finish.

01

From Prompts to Semantic Blueprints

Understand why prompts fail at scale and how semantic blueprints give AI structured, goal-driven contextual awareness.

02

Multi-Agent Orchestration With MCP

Design and orchestrate multi-agent workflows using the Model Context Protocol. Build transparent, traceable agent systems.

03

High-Fidelity RAG With Citations

Build RAG pipelines that deliver accurate, cited responses. Engineer memory systems that persist context reliably across agents.

04

The Glass-Box Context Engine

Architect a transparent, explainable context engine where every decision is traceable and debuggable in production.

05

Safeguards and Trust

Implement safeguards against prompt injection and data poisoning. Enforce trust boundaries in multi-agent environments.

06

Production Deployment and Scaling

Deploy your context-engineered system to production. Apply patterns for scaling, monitoring, and reliability.

What You Walk Away With

By the End of This Workshop You Will Have

Concrete working deliverables — not just theory and slides.

A working Glass-Box Context Engine with transparent, traceable reasoning

Multi-agent workflow orchestrated with the Model Context Protocol

High-fidelity RAG pipeline with memory and citations

Safeguards against prompt injection and data poisoning

Reusable architecture patterns for production AI systems

Certificate of completion from Packt Publishing

Your Instructor

Learn From a Bestselling AI Author With 30+ Years of Experience

Denis Rothman brings decades of production AI engineering experience to this live workshop.

Denis Rothman

Denis Rothman

Workshop Instructor · April 25, 2026

Denis Rothman is a bestselling AI author with over 30 years of experience in artificial intelligence, agent systems, and optimization. He has authored multiple cutting-edge AI books published by Packt and is renowned for making complex AI architecture concepts practical and immediately applicable. He guides you step by step through building production-ready context-engineered multi-agent systems — answering your questions live throughout the 6-hour session.

Prerequisites

Who Is This Workshop For?

Intermediate to advanced workshop. Solid Python and basic LLM experience required.

Frequently Asked Questions

Common Questions About Transparent and Explainable AI Agents

Everything you need to know before registering.

What does transparency mean for an AI agent in production? +

Transparency for a production AI agent means that for any output the system produces, you can trace exactly: which knowledge sources the agent used (RAG citations), what instructions guided its reasoning (the semantic blueprint), which other agents contributed to the output (MCP orchestration log), which safeguards were evaluated and what they found (safeguard audit trail), and what output validation was applied before delivery (validation log). This complete traceability is what the Glass-Box architecture provides.

What is explainability in AI agents and how does the Glass-Box architecture achieve it? +

Explainability means being able to produce a human-readable explanation of why the AI agent produced a specific output. The Glass-Box architecture achieves explainability by maintaining structured logs of every decision that can be rendered into natural language explanations: 'This response was generated using the following retrieved sources, guided by this semantic blueprint, in coordination with these other agents, after passing these validation checks.' The workshop covers generating these explanations from Glass-Box log data.

How does the Glass-Box architecture differ from post-hoc explainability techniques? +

Post-hoc explainability techniques like LIME and SHAP attempt to explain a model's decision after the fact by approximating the model's behavior. The Glass-Box architecture captures the actual decision process as it happens: the real context that was provided to the agent, the real retrieval results that informed the response, the real coordination steps that assembled the final output. This in-process transparency is more accurate and more actionable than post-hoc approximation techniques.

Can I make an existing black-box AI agent transparent without rebuilding it? +

Partially. You can add Glass-Box logging to the inputs and outputs of an existing agent to capture what it received and what it produced. However, the internal decision process of a black-box agent (which retrieved documents it used, how it weighted different context sources) cannot be made transparent without architectural changes. The workshop covers the incremental path from partially transparent logging to fully transparent Glass-Box architecture for existing systems.

How does Glass-Box transparency help users trust AI agent outputs? +

Glass-Box transparency builds user trust by making AI agent outputs verifiable: users can see which specific sources were cited for each claim, can verify that those sources support the claim, and can understand the reasoning process that produced the response. This verifiability transforms AI from an opaque oracle that must be trusted on faith to an accountable system whose outputs can be checked and challenged. The workshop covers how to surface Glass-Box transparency information in user-facing interfaces.

What Glass-Box transparency features are most important for enterprise AI deployments? +

Enterprise AI deployments prioritize four Glass-Box transparency features: audit trail completeness (every interaction is logged with sufficient detail for compliance review), knowledge source traceability (every factual claim can be traced to a specific retrieved source), access control logging (which users and agents accessed which knowledge resources), and safeguard evaluation records (what content moderation and safety checks were applied to each output). The workshop covers implementing all four for enterprise production deployments.

Context Engineering for Multi-Agent Systems · Cohort 2 · April 25, 2026

Ready to Build Production AI With Context Engineering?

6 hours. Bestselling AI author. Production context-engineered multi-agent system by the end. Seats are limited.

Register Now →

Saturday April 25 · 9am to 3pm EDT · Online · Packt Publishing · Cohort 2