This guide provides a step-by-step process for setting up a Kubernetes-based MLOps platform on Lambda Cloud, using the Run:AI framework. To minimize training loss, a single GPU instance may not be enough compute; therefore, a cluster of instances is recommended. The setup involves creating a head node and one or more worker nodes, installing the necessary tools such as Kubernetes, Docker, and NVIDIA driver, and configuring the cluster for use with Run:AI. The benefits of using Lambda Cloud as the underlying infrastructure include easy scaling, sharing of persistent storage across all nodes, and all nodes being equipped with GPUs.