From e152f1b0f52221e93cae5a6e4121ed3c4ecda6d0 Mon Sep 17 00:00:00 2001 From: nimlgen <138685161+nimlgen@users.noreply.github.com> Date: Mon, 26 Jan 2026 22:01:53 +0300 Subject: [PATCH] llama: use ALL2ALL (#14353) --- .../llama8b/implementations/tinybox_8xMI350X/dev_run.sh | 1 + 1 file changed, 1 insertion(+) diff --git a/examples/mlperf/training_submission_v6.0/tinycorp/benchmarks/llama8b/implementations/tinybox_8xMI350X/dev_run.sh b/examples/mlperf/training_submission_v6.0/tinycorp/benchmarks/llama8b/implementations/tinybox_8xMI350X/dev_run.sh index dab0f584a9..daa1448754 100755 --- a/examples/mlperf/training_submission_v6.0/tinycorp/benchmarks/llama8b/implementations/tinybox_8xMI350X/dev_run.sh +++ b/examples/mlperf/training_submission_v6.0/tinycorp/benchmarks/llama8b/implementations/tinybox_8xMI350X/dev_run.sh @@ -7,6 +7,7 @@ export REWRITE_STACK_LIMIT=5000000 HCQDEV_WAIT_TIMEOUT_MS=240000 export DEBUG=${DEBUG:-0} export FLASH_ATTENTION=${FLASH_ATTENTION:-1} +export ALL2ALL=${ALL2ALL:-1} export DEFAULT_FLOAT="bfloat16" OPTIM_DTYPE="bfloat16" export DP=8 BS=8 EVAL_BS=8 GRADIENT_ACC_STEPS=1