Docker Model Runner is the easiest way to run large language models locally in 2026. This live tutorial shows you how to use it to power a complete private AI assistant — connected to WhatsApp or Telegram — with no cloud dependency whatsoever.
By Packt Publishing · Refunds up to 10 days before
Docker Model Runner provides native Docker integration, an OpenAI-compatible API, and straightforward model management — making it the cleanest solution for running local LLMs in production. This tutorial shows you how to use it to its full potential.
OpenClaw is the open-source personal AI assistant that went viral in early 2026 with 200K+ GitHub stars. It runs on your own devices and connects to WhatsApp, Telegram, Slack and more. No subscription. No data leaving your machine.
Docker Model Runner is Docker's native feature for running large language models locally on your machine. It gives you an OpenAI-compatible API that OpenClaw uses as its AI brain — complete data privacy, no cloud costs.
OpenClaw gives you the assistant interface and messaging integrations. Docker Model Runner gives you the AI brain running privately on your machine. Together they create a production grade private AI assistant you fully own.
Setting this up from scattered documentation takes days of debugging. This live workshop gives you a complete guided build in 4 hours with a live instructor answering your questions. Packt has delivered 108 workshops worldwide.
Six modules. From Docker Model Runner setup to a fully deployed local LLM-powered assistant.
Understand the Gateway, channels, and skills architecture. Set up and configure OpenClaw locally from scratch.
Run and manage local LLMs using Docker Model Runner. Pull models, configure memory, and understand the OpenAI-compatible API.
Configure DM pairing, allowlists, sandbox mode, and proper access controls for your local AI deployment.
Deploy your AI assistant to real messaging platforms without sending data to any third party cloud service.
Design an extensible assistant architecture. Add skills, configure personality, and set up proactive automation.
Deploy your OpenClaw and Docker setup to a VPS for always-on availability running 24 hours a day.
A locally running LLM powering a working private AI assistant — fully your own.
A fully functional local AI assistant running on your machine
Docker Model Runner configured with your chosen LLM model
OpenClaw connected to WhatsApp or Telegram
Security and privacy configuration you can trust
A reusable architecture for future AI assistant projects
Certificate of completion from Packt Publishing
Rami Krispin is a Docker Captain with deep expertise in local LLM deployment using Docker.
Rami is a Senior Manager of Data Science and Engineering, Docker Captain, and LinkedIn Learning Instructor with deep expertise in building and deploying production AI systems. He guides you step by step from a blank terminal to a fully deployed private AI assistant — answering your questions live throughout the 4-hour session.
Developers who want to run LLMs locally and build something useful with them.
Everything you need to know about local LLM deployment with Docker Model Runner.
Docker Model Runner works with a wide range of open weight models. For a personal AI assistant use case, the instructor recommends starting with models in the 3B to 8B parameter range — they run comfortably on 16GB RAM and provide good response quality. In this tutorial you will try several models and learn to evaluate them for your specific hardware.
Local LLM inference speed depends on your hardware. On a modern laptop with 16GB RAM, you can typically expect 10 to 30 tokens per second with a 7B parameter model on CPU. The instructor covers performance expectations for different hardware configurations and model sizes during this tutorial.
Docker Model Runner significantly simplifies local LLM setup. In this tutorial you will have a local LLM running within the first hour of the 4-hour session. The instructor walks through every step including Docker setup, model pulling, and API configuration.
Modern open weight models have improved dramatically. For personal assistant tasks — answering questions, summarising content, helping with writing, and conversational tasks — models available through Docker Model Runner perform very well. The instructor covers model selection to help you choose the best fit for your use case.
Docker Model Runner makes switching between local LLMs straightforward. This tutorial covers how to pull multiple models, how to configure OpenClaw to use a specific model, and how to switch models without disrupting your assistant setup.
Yes. All local LLM inference through Docker Model Runner happens entirely on your own machine. No conversation data is sent to any external server. Your private AI assistant powered by a local LLM is completely air-gapped from cloud AI providers.
4 hours. Live Docker Captain instructor. Running local LLM by the end. Seats are limited.
Register Now →Sunday April 26 · 9am to 1pm EDT · Online · Packt Publishing