Update PyTorch compatibility documentation with PyTorch2.9 for ROCm7.1.1

This commit is contained in:
Jithun Nair
2025-11-19 19:15:04 -06:00
committed by GitHub
2 changed files with 25 additions and 0 deletions

View File

@@ -79,6 +79,7 @@ CX
Cavium Cavium
CentOS CentOS
ChatGPT ChatGPT
Cholesky
CoRR CoRR
Codespaces Codespaces
Commitizen Commitizen
@@ -270,6 +271,7 @@ LLM
LLMs LLMs
LLVM LLVM
LM LM
logsumexp
LRU LRU
LSAN LSAN
LSan LSan
@@ -322,6 +324,7 @@ Mooncake
Mpops Mpops
Multicore Multicore
Multithreaded Multithreaded
mx
MXFP MXFP
MyEnvironment MyEnvironment
MyST MyST

View File

@@ -399,6 +399,28 @@ with ROCm.
**Note:** Only official release exists. **Note:** Only official release exists.
Key features and enhancements for PyTorch 2.9 with ROCm 7.1.1
================================================================================
- Scaled Dot Product Attention (SDPA) upgraded to use AOTriton version 0.11b
- Default hipBLASLt support enabled for gfx908 architecture on ROCm 6.3 and later
- MIOpen now supports channels last memory format for 3D convolutions and batch normalization
- NHWC convolution operations in MIOpen optimized by eliminating unnecessary transpose operations
- Improved tensor.item() performance by removing redundant synchronization
- Enhanced performance for element-wise operations and reduction kernels
- Added support for grouped GEMM operations through fbgemm_gpu generative AI components
- Resolved device error in Inductor when using CUDA graph trees with HIP
- Corrected logsumexp scaling in AOTriton-based SDPA implementation
- Added stream graph capture status validation in memory copy synchronization functions
Key features and enhancements for PyTorch 2.8 with ROCm 7.1 Key features and enhancements for PyTorch 2.8 with ROCm 7.1
================================================================================ ================================================================================