From 0ddc7d167b6eba2b91dad028060320de0aa8aeca Mon Sep 17 00:00:00 2001 From: ninggu Date: Tue, 24 Feb 2026 16:43:55 -0800 Subject: [PATCH] Update llama3.1-405b 64 node README.md --- .../nemo-pretraining-gke/64node-FP8CS-GBS2048/recipe/README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/training/a4x/llama3-1-405b/nemo-pretraining-gke/64node-FP8CS-GBS2048/recipe/README.md b/training/a4x/llama3-1-405b/nemo-pretraining-gke/64node-FP8CS-GBS2048/recipe/README.md index fead3a3f..84ca143b 100644 --- a/training/a4x/llama3-1-405b/nemo-pretraining-gke/64node-FP8CS-GBS2048/recipe/README.md +++ b/training/a4x/llama3-1-405b/nemo-pretraining-gke/64node-FP8CS-GBS2048/recipe/README.md @@ -83,7 +83,7 @@ gcloud container clusters get-credentials $CLUSTER_NAME --region $CLUSTER_REGION ### Configure and submit a pretraining job -#### Using 16 node (64 gpus) fp8 precision +#### Using 64 node (256 gpus) fp8 precision To execute the job with the default settings, run the following command from your client: