By watching this webinar, replay you'll learn some of the common challenges with scaling-out deep learning training and inference deployment on data centers and public cloud using Kubernetes on NVIDIA GPUs. Through examples, we'll review a typical workflow for AI deployments on Kubernetes. We'll discuss advanced deployment options such as deploying to heterogenous GPU clusters, specifying GPU memory requirements, and analyzing and monitoring GPU utilizations using NVIDIA DCGM, Prometheus and Grafana.