Peter Park
548d31f990
fix broken image in megatron-lm-v24.12-dev.rst ( #5043 )
2025-07-15 10:57:12 -04:00
Pratik Basyal
544186aef8
ROCm for HPC table update for Develop ( #5015 ) ( #5016 ) ( #5019 )
...
* ROCm for HPC table update for 6.4.0 (#5015 ) (#5016 )
* 6.4.0 updates synced
* Minor change
* Link update
2025-07-09 14:57:53 -04:00
Peter Park
22524eeaa5
fix xrefs in vllm-0.9.0.1-20250605.rst ( #5017 )
2025-07-09 14:38:24 -04:00
Peter Park
d471b04cd5
Update vLLM Docker doc for 07/02
2025-07-09 11:38:27 -04:00
Peter Park
3b3fc4894b
Fix xrefs and Sphinx warnings in documentation
...
Fix xrefs and Sphinx warnings in documentation
2025-07-08 13:22:53 -04:00
Peter Park
58b3ad0509
Fix Docker run commands in Megatron-LM Docker doc ( #4996 )
...
* fix megatron-lm docker run commands
* update --shm-size option
2025-07-02 14:19:27 -04:00
Peter Park
d0c8ba0805
Add Wan2.1 to PyTorch inference Docker documentation ( #4984 )
...
* add wan2.1 to pyt inference models
* update group name
* fix container tag
* fix group name
* change documented data type to bfloat16
* fix col width
2025-07-02 09:58:37 -04:00
Peter Park
2196fc9a2f
Fix pytorch training 25.6 doc ( #4956 )
...
* fix pytorch-training history
* fix pytorch-training
fix
2025-06-23 13:45:50 -04:00
Peter Park
91a541f8b9
Update PyTorch training benchmark doc for v25.6 ( #4950 )
...
* update pytorch-training docker details
* add previous version
* add models data
* update models data id
* add models picker
* update data
* update fmt
fmt
* update data yaml
* update template
* update data
* fix
* fix vllm-0.6.4 broken link
* fix vllm history
2025-06-23 09:26:15 -04:00
Peter Park
34f8d57ece
Organize version histories in ROCm for AI benchmark Docker docs ( #4948 )
...
* add vllm 0.8.3 20250415
update prev versions table
* add vllm previous versions page
* move index to vllm-history
* add standalone megatron-lm version history
* add pytorch training version history
* fix
* add vllm-0.4.3
* add vllm-0.6.4
* update vllm-history
* add vllm-0.7.3
* add vllm-0.6.6
* add notes
* fix vllm readme links
fix main page link
* add latest version to previous versions list
* add jax-maxtext history
* fix jax-maxtext history
* add pytorch-training history
* add link in jax-maxtext 25.4
* add megatron-lm history
* fix datatemplate path for vllm 0.8.3
* fix jax-maxtext history link
* update note about performance measurements
* add vllm 0.8.5_20250521 previous version
* consistency fixes
2025-06-20 15:01:38 -04:00
yugang-amd
55f95adc7c
Update for vllm -06/10 ( #4943 )
2025-06-20 08:41:37 -04:00
yugang-amd
7b7eaf69f2
remove broken xref ( #4939 )
2025-06-18 10:15:53 -04:00
Peter Park
d69037bfcc
Fix Sphinx issue in vllm-benchmark 0.8.5-20250513 previous version ( #4924 )
...
* fix sphinx issue in vllm-benchmark 0.8.5-20250513 previous version
* update article_info in conf.py
* update rocm/vllm
2025-06-13 15:03:51 -04:00
Peter Park
cfb3504d77
Add Mochi Video to pytorch-inference-benchmark-models.yaml
...
Add Mochi Video to pytorch-inference-benchmark-models.yaml
2025-06-10 13:18:41 -04:00
yugang-amd
830f2d5edf
Update for vllm -05/27 ( #4886 )
...
* Update vLLM inference benchmark Docker page for rocm/vllm 5/27
* update repo for Pytorch
2025-06-05 13:30:20 -04:00
yugang-amd
53d3e092d3
Fix broken link ( #4854 )
2025-05-31 13:01:34 -04:00
Peter Park
2eb8bf4963
Fix typo in Megatron-LM Docker pull tags ( #4829 )
2025-05-28 15:18:00 -04:00
Peter Park
cebf0f5975
Add latest rocm/vllm Docker details in vLLM inference benchmark guide ( #4824 )
...
* update rocm/vllm Docker details to latest release
* Add previous vLLM version
* fix 'further reading' xrefs
* improve model grouping names
* fix links
* update model picker text
2025-05-28 14:20:18 -04:00
Peter Park
505041d90a
Document specs for Radeon RX 9070 + small fix in megatron-lm doc ( #4780 )
...
* Document specs for Radeon RX 9070
* fix wrong version in megatron-lm.rst
2025-05-22 16:28:17 -04:00
Peter Park
9ed65a81c4
Add Megatron-LM benchmark doc 5/2 ( #4778 )
...
* reorg files
* add tabs
* update template
* update template
* update wordlist and toc
* add previous version to doc
* add selector paragraph
* update wordlist.txt
2025-05-22 14:28:18 -04:00
Peter Park
0a77e7b3a5
docs: Add system health check doc under ROCm for AI ( #4736 )
...
* add initial draft
* add to toc and install page
* update wording
* improve documentation structure
* resturcture and expand content
* add to training section
* add to conf.py article_pages
* Update docs/how-to/rocm-for-ai/includes/system-health-benchmarks.rst
Co-authored-by: Leo Paoletti <164940351+lpaoletti@users.noreply.github.com >
* Update docs/how-to/rocm-for-ai/includes/system-health-benchmarks.rst
Co-authored-by: Leo Paoletti <164940351+lpaoletti@users.noreply.github.com >
* update wordlist.txt
* Update docs/how-to/rocm-for-ai/includes/system-health-benchmarks.rst
Co-authored-by: Leo Paoletti <164940351+lpaoletti@users.noreply.github.com >
* inference --> AI workloads
* udpate toc
* update article_pages in conf.py
* Update system validation notes in training docs
* fix links in prerequisite-system-validation
* wording
* add note
* consistency
* remove extra files
* fix links
* add links to training index page
---------
Co-authored-by: Leo Paoletti <164940351+lpaoletti@users.noreply.github.com >
2025-05-13 15:54:48 -04:00
Peter Park
90a651d2b6
Merge pull request #4725 from peterjunpark/docs/quark-model-quantization
...
Add quark in model-quantization.rst
2025-05-08 10:34:39 -04:00
Peter Park
bb7af3351a
Fix incorrect throughput benchmark command in inference/vllm-benchmark.rst ( #4723 )
...
* update inference index to include pyt inference
* fix incorrect command in throughput benchmark
* wording
2025-05-08 09:24:51 -04:00
Peter Park
186c281aba
fix links in pytorch-inference-benchmark.rst ( #4713 )
2025-05-06 13:34:55 -04:00
Peter Park
d44ea40a0d
Add MPT-30B + LLM Foundry doc ( #4704 )
...
* add mpt-30b doc
* add tunableop note
* update MPT doc
* add section
* update wordlist
* fix flash attention version
* update "applies to"
* address review feedback
* Update docs/how-to/rocm-for-ai/training/benchmark-docker/mpt-llm-foundry.rst
Co-authored-by: Leo Paoletti <164940351+lpaoletti@users.noreply.github.com >
* Update docs/how-to/rocm-for-ai/training/benchmark-docker/mpt-llm-foundry.rst
Co-authored-by: Leo Paoletti <164940351+lpaoletti@users.noreply.github.com >
* Update docs/how-to/rocm-for-ai/training/benchmark-docker/mpt-llm-foundry.rst
Co-authored-by: Leo Paoletti <164940351+lpaoletti@users.noreply.github.com >
* update docker details to pytorch-training-v25.5
* update
---------
Co-authored-by: Leo Paoletti <164940351+lpaoletti@users.noreply.github.com >
2025-05-02 12:13:20 -04:00
Peter Park
7458fcb7ab
Update JAX MaxText benchmark doc to v25.5 ( #4695 )
...
* fix shell cmd formatting
* add previous versions section
* update docker details and add llama 3.3
* update missed docker image tags to 25.5
2025-04-28 17:52:53 -04:00
Peter Park
16d6e59003
fix link to pytorch-training v25.4 doc ( #4696 )
2025-04-28 17:52:33 -04:00
Peter Park
a66bc1d85e
fix link to previous version in vllm-benchmark.rst ( #4689 )
2025-04-24 17:54:04 -04:00
Peter Park
36b6ffaf7c
Add QwQ 32B to vllm-benchmark.rst ( #4685 )
...
* Add Qwen2 MoE 2.7B to vllm-benchmark-models.yaml
* Add QwQ-32B-Preview to vllm-benchmark-models.yaml
* add links to performance results
words
* change "performance validation" to "performance testing"
* remove "-Preview" from QwQ-32B
* move qwen2 MoE after qwen2
* add TunableOp section
* fix formatting
* add link to TunableOp doc
* add tunableop note
* fix vllm-benchmark template
* remove cmdline option for --tunableop on
* update docker details
* remove "training"
* remove qwen2
2025-04-24 16:44:34 -04:00
Peter Park
40e4ba3ecc
Update vLLM inference benchmark Docker guide ( #4653 )
...
* Remove JAIS 13B and 30B
* update Docker details - vLLM 0.8.3
* add previous version
* Update docs/how-to/rocm-for-ai/inference/vllm-benchmark.rst
* fix link to previous version
2025-04-24 15:59:13 -04:00
Peter Park
1f41ce26be
Add note for chai-1 benchmark Docker in pytorch-inference-benchmark.rst ( #4684 )
2025-04-24 15:48:53 -04:00
Peter Park
c3faa9670b
Add PyTorch inference benchmark Docker guide (+ CLIP and Chai-1) ( #4654 )
...
* update vLLM links in deploy-your-model.rst
* add pytorch inference benchmark doc
* update toc and vLLM title
* remove previous versions
* update
* wording
* fix link and "applies to"
* add pytorch to wordlist
* add tunableop note to clip
* make tunableop note appear to all models
* Update docs/how-to/rocm-for-ai/inference/pytorch-inference-benchmark.rst
Co-authored-by: Leo Paoletti <164940351+lpaoletti@users.noreply.github.com >
* Update docs/how-to/rocm-for-ai/inference/pytorch-inference-benchmark.rst
Co-authored-by: Leo Paoletti <164940351+lpaoletti@users.noreply.github.com >
* Update docs/how-to/rocm-for-ai/inference/pytorch-inference-benchmark.rst
Co-authored-by: Leo Paoletti <164940351+lpaoletti@users.noreply.github.com >
* Update docs/how-to/rocm-for-ai/inference/pytorch-inference-benchmark.rst
Co-authored-by: Leo Paoletti <164940351+lpaoletti@users.noreply.github.com >
* fix incorrect links
* wording
* fix wrong docker pull tag
---------
Co-authored-by: Leo Paoletti <164940351+lpaoletti@users.noreply.github.com >
2025-04-23 17:35:52 -04:00
Peter Park
9ff3c2c885
Update PyTorch training Docker doc for 25.5 ( #4638 )
...
* update pytorch-training to 25.5
* remove llama 2
* Revert "remove llama 2"
This reverts commit dab672fa7bcbd8bff730382c14177df4301a537d.
* add previous version
* fix run cmd
* add link to docker hub
* fix linting issue
* add Llama 3.3 70B
* update
2025-04-15 18:16:22 -04:00
Peter Park
d057d49af1
Fix vllm Dockerfile.rocm path ( #4628 )
2025-04-15 11:26:54 -04:00
Peter Park
310864e653
fix link to Dockerfile.rocm ( #4573 )
2025-04-14 10:10:03 -04:00
Parag Bhandari
493585dfbb
Merge branch 'develop' of github.com:ROCm/ROCm into develop
2025-04-11 15:15:43 -04:00
Parag Bhandari
e756d99f65
Merge branch 'develop-internal' into develop
2025-04-11 15:15:19 -04:00
Pratik Basyal
686fcece1d
PRE GA Day 640 update for resetting link and HPC application list ( #367 )
...
* Links reset to point to latest from stg, internal, RTD, and develop
* ROCm for HPC updated
* GA prep changes
2025-04-11 14:12:57 -05:00
pbhandar-amd
131e34f582
Update w6000-v620.md
2025-04-11 15:11:34 -04:00
Parag Bhandari
db3c46fccf
Merge branch 'develop-internal' into develop
2025-04-11 14:32:09 -04:00
Dominic Widdows
715cce53de
Update workload.rst with small export fix ( #4425 )
...
Tiny fix that removes the "export" directive.
` export HIP_FORCE_DEV_KERNARG=1 hipblaslt-bench ...`
leads to
bash: export: `hipblaslt-bench': not a valid identifier
whereas just starting with HIP_FORCE_DEV_KERNARG=1 passes this env var to the hipblaslt-bench process, which I think is the intention here.
2025-04-03 13:01:26 -04:00
Peter Park
ea66bf386a
Fix more links in documentation ( #4551 )
...
* fix vllm engine args link
* remove RDNA subtree in under system optimization in toc
* fix RDNA 2 architecture PDF link
* fix CLR LICENSE.txt link
* fix rocPyDecode license link
2025-04-01 15:56:34 -04:00
Peter Park
ac2c5e72d4
Fix links in documentation
2025-04-01 15:39:20 -04:00
Peter Park
424e6148bd
Add MaxText training Docker doc
...
Add MaxText training Docker doc
2025-03-28 11:25:06 -04:00
Pratik Basyal
a0faccba37
AMD GPU Docs System optimization migration changes in ROCm Docs Develop ( #4538 )
...
* AMD GPU Docs System optimization migration changes in ROCm Docs (#296 )
* System optimization migration changes in ROCm
* Linting issue fixed
* Linking corrected
* Minor change
* Link updated to Instinct.docs.amd.com
* ROCm docs grid updated by removing IOMMU.rst, pcie-atomics, and oversubscription pages
* Files removed and reference fixed
* Reference text updated
* GPU atomics from 6.4.0 removed
2025-03-27 16:38:10 -04:00
Pratik Basyal
544149631a
AMD GPU Docs System optimization migration changes in ROCm Docs ( #296 )
...
* System optimization migration changes in ROCm
* Linting issue fixed
* Linking corrected
* Minor change
* Link updated to Instinct.docs.amd.com
* ROCm docs grid updated by removing IOMMU.rst, pcie-atomics, and oversubscription pages
* Files removed and reference fixed
* Reference text updated
2025-03-26 10:01:33 -04:00
Peter Park
58d42ec50b
Improve "tuning guides" landing page ( #4504 )
...
* Improve "tuning guides" landing page
* Update docs/how-to/gpu-performance/mi300x.rst
Co-authored-by: Pratik Basyal <pratik.basyal@amd.com >
* Update docs/how-to/gpu-performance/mi300x.rst
Co-authored-by: Pratik Basyal <pratik.basyal@amd.com >
* change tuning to optimization
---------
Co-authored-by: Pratik Basyal <pratik.basyal@amd.com >
2025-03-25 13:54:27 -04:00
Peter Park
8f359da39e
Update Megatron-LM doc for 25.4 ( #4520 )
...
* update megatron-lm doc
* update 'previous versions'
* add missing space
* update docker pull tag
* Update options and docker pull tag
* Add performance measurements link to megatron-lm doc
* fix previous versions table
* words
* Simplify system validation section
* minor fixes
* fix perv versions tbl
2025-03-21 16:49:55 -04:00
Peter Park
2fca094531
PyTorch training Docker update 25.4 ( #4482 )
...
* remove orphan tag
* add hugging face PEFT
* update "previous versions"
* data == ultrachat 200k
* fix "llama 2"
* add ultrachat to wordlist
* fix previous versions table
* add performance measurements
* add mi325x
* fix prev version
* change 'validation' to 'testing
* fix dir name
* fix backtick
2025-03-13 13:40:00 -04:00
Peter Park
9b2ce2b634
Update vLLM performance Docker docs ( #4491 )
...
* add links to performance results
words
* change "performance validation" to "performance testing"
* update vLLM docker 3/11
* add previous versions
add previous versions
* fix llama 3.1 8b model repo name
* words
2025-03-13 10:04:21 -04:00