randyh62
2b83a962a0
Use intersphinx links for deep learning ( #5859 )
...
* Use intersphinx links for deep learning
* Update deep-learning-rocm.rst
remove Taichi
* Update deep-learning-rocm.rst
Change Install link to "link"
* Apply suggestion from @randyh62
OK
2026-01-20 09:17:37 -08:00
peterjunpark
4347a11bc4
Doc update for vLLM refactor #5855 ( #5856 )
...
(cherry picked from commit a745e45dcb )
2026-01-15 11:34:02 -05:00
Swati Rawat
4184d1ee1f
Update docs/how-to/rocm-for-ai/training/benchmark-docker/previous-versions/megatron-lm-v24.12-dev.rst
...
Co-authored-by: peterjunpark <git@peterjunpark.com >
2026-01-08 16:46:22 +05:30
Swati Rawat
0786c328c1
Update docs/how-to/rocm-for-ai/system-setup/prerequisite-system-validation.rst
...
Co-authored-by: peterjunpark <git@peterjunpark.com >
2026-01-08 16:46:22 +05:30
Swati Rawat
88ea6072f5
Update docs/how-to/rocm-for-ai/system-setup/prerequisite-system-validation.rst
...
Co-authored-by: peterjunpark <git@peterjunpark.com >
2026-01-08 16:46:22 +05:30
Swati Rawat
b32dcc8570
Update docs/how-to/rocm-for-ai/system-setup/prerequisite-system-validation.rst
...
Co-authored-by: peterjunpark <git@peterjunpark.com >
2026-01-08 16:46:22 +05:30
Swati Rawat
0faa92e922
Update docs/how-to/rocm-for-ai/training/benchmark-docker/previous-versions/megatron-lm-v24.12-dev.rst
...
Co-authored-by: peterjunpark <git@peterjunpark.com >
2026-01-08 16:46:21 +05:30
Swati Rawat
26ae989602
Update docs/how-to/rocm-for-ai/training/benchmark-docker/previous-versions/megatron-lm-v24.12-dev.rst
...
Co-authored-by: peterjunpark <git@peterjunpark.com >
2026-01-08 16:46:21 +05:30
srawat
4402dc4147
Update single-gpu-fine-tuning-and-inference.rst
2026-01-08 16:46:21 +05:30
srawat
5eda438e0a
Update multi-gpu-fine-tuning-and-inference.rst
2026-01-08 16:46:20 +05:30
srawat
049784e1a7
Update prerequisite-system-validation.rst
2026-01-08 16:42:18 +05:30
srawat
f12169c5b7
replace rocm-smi reference with amd-smi
2026-01-08 16:42:18 +05:30
peterjunpark
b35d1a0627
fix(primus-pytorch.rst): FP8 config instead of BF16 ( #5839 )
...
(cherry picked from commit 2dc22ca890 )
2026-01-07 13:51:50 -05:00
peterjunpark
7d2feaa8b1
Fix inconsistency in xDiT doc ( #5823 )
...
Fix inconsistency in xDiT doc
(cherry picked from commit 172b0f7c08 )
2025-12-29 10:29:59 -05:00
peterjunpark
2a65394e32
Update docs for xDiT diffusion inference 25.13 Docker release ( #5820 )
...
* archive previous version
* add xdit 25.13
* update history index
* add perf results section
(cherry picked from commit c67fac78bd )
2025-12-29 08:45:29 -05:00
peterjunpark
268c1332c9
Update training docs for Primus/25.11 ( #5819 )
...
* update conf and toc.yml.in
* archive previous versions
archive data files
update anchors
* primus pytorch: remove training batch size args
* update primus megatron run cmds
multi-node
* update primus pytorch
update
* update
update
* update docker tag
(cherry picked from commit e0b8ec4dfb )
2025-12-29 08:45:17 -05:00
peterjunpark
512e311041
Update xdit diffusion inference history ( #5808 ) ( #5809 )
...
* Update xdit diffusion inference history
* fix
(cherry picked from commit 3a43bacdda )
2025-12-22 11:14:57 -05:00
peterjunpark
ad4f486635
fix link to ROCm PyT docker image ( #5803 ) ( #5804 )
...
(cherry picked from commit 48d8fe139b )
2025-12-19 15:51:20 -05:00
peterjunpark
485886712b
clean up formatting in FA2 page ( #5795 ) ( #5796 )
...
(cherry picked from commit 7455fe57b8 )
2025-12-19 09:38:20 -05:00
peterjunpark
1cd6a14a22
Update Flash Attention guidance in "Model acceleration libraries" ( #5793 )
...
* flash attention update
Signed-off-by: seungrok.jung <seungrok.jung@amd.com >
flash attention update
Signed-off-by: seungrok.jung <seungrok.jung@amd.com >
flash attention update
Signed-off-by: seungrok.jung <seungrok.jung@amd.com >
sentence-case heading
* Update docs/how-to/rocm-for-ai/inference-optimization/model-acceleration-libraries.rst
Co-authored-by: Leo Paoletti <164940351+lpaoletti@users.noreply.github.com >
* Apply suggestions from code review
Co-authored-by: Leo Paoletti <164940351+lpaoletti@users.noreply.github.com >
---------
Co-authored-by: seungrok.jung <seungrok.jung@amd.com >
Co-authored-by: Leo Paoletti <164940351+lpaoletti@users.noreply.github.com >
(cherry picked from commit 52c0a47e84 )
2025-12-19 09:00:40 -05:00
peterjunpark
a17f04a3b5
Update documentation for JAX training MaxText 25.11 release ( #5789 ) ( #5790 )
...
(cherry picked from commit cbab9a465d )
2025-12-18 11:26:42 -05:00
peterjunpark
94de66ef3f
[docs/7.1.1] Publish vLLM and xDiT doc updates ( #5787 )
...
* vLLM inference benchmark 1210 (#5776 )
* Archive previous ver
fix anchors
* Update vllm.rst and data yaml for 20251210
(cherry picked from commit 1b4f25733d )
* xDiT diffusion inference v25.12 documentation update (#5786 )
* Add xdit-diffusion ROCm docs page.
* Update template formatting and fix sphinx warnings
* Add System Validation section.
* Add sw component versions/commits.
* Update to use latest v25.10 image instead of v25.9
* Update commands and add FLUX instructions.
* Update Flux instructions. Change image tag. Describe as diffusion inference instead of specifically video.
* git rm xdit-video-diffusion.rst
* Docs for v25.12
* Add hyperlinks to components
* Command fixes
* -Diffusers suffix
* Simplify yaml file and cleanup main rst page.
* Spelling, added 'js'
* fix merge conflict
fix
---------
Co-authored-by: Kristoffer <kristoffer.torp@amd.com >
(cherry picked from commit 459283da3c )
---------
Co-authored-by: Kristoffer <kristoffer.torp@amd.com >
2025-12-17 10:28:30 -05:00
Pratik Basyal
e5cebe7b4e
Taichi removed from ROCm docs [Develop] ( #5779 ) ( #5781 )
...
* Taichi removed from ROCm docs
* Warnings fixed
2025-12-16 13:24:12 -05:00
Matt Williams
0d17c96f7f
Fixing link redirects ( #5758 )
...
* Update multi-gpu-fine-tuning-and-inference.rst
* Update pytorch-training-v25.6.rst
* Update pytorch-compatibility.rst
2025-12-10 11:31:26 -05:00
peterjunpark
8f45b791fe
Fix Primus PyTorch doc: training.batch_size -> training.local_batch_size ( #5748 ) ( #5749 )
...
(cherry picked from commit bf74351e5a )
2025-12-08 13:59:00 -05:00
yugang-amd
f7c7587b10
xdit-diffusion v25.11 docs ( #5743 )
2025-12-05 17:08:21 -05:00
peterjunpark
d6d4d2ef92
fix docker hub links for primus:v25.10 ( #5738 )
...
(cherry picked from commit 453751a86f )
2025-12-04 09:21:53 -05:00
peterjunpark
8647ebcf76
Update training Docker docs for Primus 25.10 ( #5737 )
...
(cherry picked from commit fb644412d5 )
2025-12-04 09:21:53 -05:00
Alex Xu
007f24fe7b
Merge remote-tracking branch 'external/develop' into sync-develop-from-external
2025-11-26 10:09:04 -05:00
Pratik Basyal
fb098b6354
Initial changes for 7.1.1 release notes ( #622 )
...
* Changelog and tables updates for 7.1.1 release notes
* Changelog synced
* Naming udpated
* Added upcoming changes for composable kernel
* Update RELEASE.md
Co-authored-by: Pratik Basyal <prbasyal@amd.com >
* Update RELEASE.md
* Highlights udpated for DGL, ROCm-DS, and HIP documentation
* Changelog synced"
* Offline, runfile and ROCm Bandwidth test updated
* CK/AITER highlight added
* Changelog synced
* AI model highlight updated
* PLDM version added
* Changelog updated
* Leo's feedback incorporated
* Compatibility and PLDM versions udpated
* New docs update added
* ROCm resolved issue added
* Review feedback added
* Link added
* PLDM updated
* PLDM table udpated
* Changes
---------
Co-authored-by: spolifroni-amd <Sandra.Polifroni@amd.com >
2025-11-17 12:09:59 -05:00
peterjunpark
eb956cfc5c
Fixed wording related to VLLM_V1_USE_PREFILL_DECODE_ATTENTION ( #5605 )
...
Co-authored-by: Hongxia Yang <hongxia.yang@amd.com >
2025-11-11 09:22:11 -05:00
peterjunpark
e05cdca54f
Fix references to vLLM docs ( #5651 )
2025-11-11 09:00:07 -05:00
anisha-amd
04c7374f41
Docs: frameworks 25.10 - compatibility - DGL and llama.cpp ( #5648 )
2025-11-10 15:26:54 -05:00
yugang-amd
674dc355e4
vLLM 10/24 release ( #5626 )
...
* vLLM 10/24 release
* updates per SME inputs
* Update docs/how-to/rocm-for-ai/inference/benchmark-docker/vllm.rst
Co-authored-by: Jeffrey Novotny <jnovotny@amd.com >
---------
Co-authored-by: Jeffrey Novotny <jnovotny@amd.com >
2025-11-05 11:13:50 -05:00
peterjunpark
1515fb3779
Revert "Add xdit diffusion docs ( #5576 )" ( #5580 )
...
This reverts commit 4132a2609c .
2025-10-27 16:22:28 -04:00
Kristoffer
4132a2609c
Add xdit diffusion docs ( #5576 )
...
* Add xdit video diffusion base page.
* Update supported accelerators.
* Remove dependency on mad-tags.
* Update docker pull section.
* Update container launch instructions.
* Improve launch instruction options and layout.
* Add benchmark result outputs.
* Fix wrong HunyuanVideo path
* Finalize instructions.
* Consistent title.
* Make page and side-bar titles the same.
* Updated wordlist. Removed note container reg HF.
* Remove fp8_gemms in command and add release notes.
* Update accelerators naming.
* Add note regarding OOB performance.
* Fix admonition box.
* Overall fixes.
2025-10-27 14:56:55 +01:00
peterjunpark
35ca027aa4
Fix broken links under rocm-for-ai/ ( #5564 )
2025-10-23 14:39:58 -04:00
peterjunpark
90c1d9068f
add xref to vllm v1 optimization guide in workload.rst ( #5560 )
2025-10-22 13:47:46 -04:00
peterjunpark
cb8d21a0df
Updates to the vLLM optimization guide for MI300X/MI355X ( #5554 )
...
* Expand vLLM optimization guide for MI300X/MI355X with comprehensive AITER coverage. attention backend selection, environment variables (HIP/RCCL/Quick Reduce), parallelism strategies, quantization (FP8/FP4), engine tuning, CUDA graph modes, and multi-node scaling.
Co-authored-by: PinSiang <pinsiang.tan@embeddedllm.com >
Co-authored-by: Hongxia Yang <62075498+hongxiayang@users.noreply.github.com >
Co-authored-by: pinsiangamd <pinsiang.tan@amd.com >
Co-authored-by: Jeffrey Novotny <jnovotny@amd.com >
2025-10-22 12:54:25 -04:00
peterjunpark
a613bd6824
JAX Maxtext v25.9 doc update ( #5532 )
...
* archive previous version (25.7)
* update docker components list for 25.9
* update template
* update docker pull tag
* update
* fix intro
2025-10-17 11:31:06 -04:00
peterjunpark
14bb59fca9
Update Megatron/PyTorch Primus 25.9 docs ( #5528 )
...
* add previous versions
* Fix heading levels in pages using embedded templates (#5468 )
* update primus-megatron doc
update megatron-lm doc
update templates
fix tab
update primus-megatron model configs
Update primus-pytorch model configs
fix css class
add posttrain to pytorch-training template
update data sheets
update
update
update
update docker tags
* Add known issue and update Primus/Turbo versions
* add primus ver to histories
* update primus ver to 0.1.1
* fix leftovers from merge conflict
2025-10-16 12:51:30 -04:00
anisha-amd
a98236a4e3
Main Docs: references of accelerator removal and change to GPU ( #5495 )
...
* Docs: references of accelerator removal and change to GPU
Co-authored-by: Leo Paoletti <164940351+lpaoletti@users.noreply.github.com >
Co-authored-by: Pratik Basyal <pratik.basyal@amd.com >
2025-10-16 11:22:10 -04:00
Pratik Basyal
036aaa2e78
ROCm for HPC topic updated Develop ( #5504 )
...
* ROCm for HPC topic updated
* ROCm for HPC topic udpated
* Minor editorial
2025-10-10 22:31:51 -04:00
peterjunpark
68e8453ca5
Update vLLM doc for 10/6 release and bump rocm-docs-core to 1.26.0 ( #5481 )
...
* archive previous doc version
* update model/docker data and doc templates
* Update "Reproducing the Docker image"
* fix: truncated commit hash doesn't work for some reason
* bump rocm-docs-core to 1.26.0
* fix numbering
fix
* update docker tag
* update .wordlist.txt
2025-10-08 16:23:40 -04:00
peterjunpark
eeea0d2180
Fix heading levels in pages using embedded templates ( #5468 )
2025-10-03 13:33:14 -04:00
anisha-amd
93c6d17922
Docs: frameworks 25.09 - compatibility - FlashInfer and llama.cpp ( #5462 )
2025-10-02 13:51:36 -04:00
peterjunpark
2e1b4dd5ee
Add multi-node setup instructions for training perf Dockers ( #5449 )
...
---------
Co-authored-by: Jeffrey Novotny <jnovotny@amd.com >
2025-09-30 14:53:38 -04:00
Peter Park
fd59b5fbac
fix links in docs ( #5446 )
2025-09-29 15:27:32 -04:00
Pratik Basyal
d92d9268dc
Use of Radeon and Ryzen reference updated [Develop] ( #5432 )
...
* Use of Radeon and Ryzen reference updated
* Pytorch link update
2025-09-24 19:07:41 -05:00
Peter Park
442d7e4750
Add env var note to vllm.rst for MoE models and fix links in docs ( #5415 )
...
* docs(vllm.rst): add performance note for MoE models
* docs: fix links
update vllm readme link 20250521
fix links
2025-09-22 15:58:43 -04:00