Ship custom ML and generative models without wrestling with DevOps. In this hands-on workshop, you’ll use Modal — a serverless compute platform for AI and data workloads — to define infrastructure in Python, run on-demand CPU/GPU hardware, and scale from one request to hundreds of workers in seconds. You’ll learn Modal’s core concepts, compare CPU vs. GPU execution, explore parallelism patterns, and deploy a production-style Stable Diffusion inference pipeline.