From db2f7bab82c57d6ddb02455a39deafc3e005e6a6 Mon Sep 17 00:00:00 2001 From: Matteo Bunino <48362942+matbun@users.noreply.github.com> Date: Tue, 7 Jan 2025 16:39:52 +0100 Subject: [PATCH] Update README.md --- tutorials/distributed-ml/torch-kubeflow-1/README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/tutorials/distributed-ml/torch-kubeflow-1/README.md b/tutorials/distributed-ml/torch-kubeflow-1/README.md index 1e484a28..8d12f4d2 100644 --- a/tutorials/distributed-ml/torch-kubeflow-1/README.md +++ b/tutorials/distributed-ml/torch-kubeflow-1/README.md @@ -15,7 +15,7 @@ for more details. ## Installing Kubeflow's Training Operator First, install the [training operator](https://www.kubeflow.org/docs/components/training/installation/). -Python SDK are not needed for this tutorial. +Python SDK is not needed for this tutorial. Example for `v1.8.1`: @@ -74,7 +74,7 @@ first. To manage both Master and Worker pods use: kubectl describe pod torchrun-cpu-worker-0 -n kubeflow kubectl describe pod torchrun-cpu-master-0 -n kubeflow -# Get the logs from the commands run the pods +# Get the logs from the pods kubectl logs torchrun-cpu-master-0 -n kubeflow kubectl logs torchrun-cpu-worker-0 -n kubeflow ```