diff --git a/docs/benchmark-docker/inference-vllm-gpt-oss-120b.rst b/docs/benchmark-docker/inference-vllm-gpt-oss-120b.rst index 1644b1c95..be592d999 100644 --- a/docs/benchmark-docker/inference-vllm-gpt-oss-120b.rst +++ b/docs/benchmark-docker/inference-vllm-gpt-oss-120b.rst @@ -5,8 +5,7 @@ Benchmark GPT OSS 120B inference with vLLM This section provides instructions to test the inference performance of OpenAI GPT OSS 120B on the vLLM inference engine. The accompanying Docker image integrates `ROCm 7.0 `__ with vLLM, and is tailored for AMD Instinct -MI355X, MI350X, and MI300X series accelerators. This benchmark does not support other -GPUs. +MI355X and MI350X accelerators. Follow these steps to pull the required image, spin up the container with the appropriate options, download the model, and run the throughput test.