mirror of
https://github.com/ROCm/ROCm.git
synced 2026-01-07 22:03:58 -05:00
clean up formatting in FA2 page (#5795)
This commit is contained in:
@@ -41,16 +41,16 @@ Installing Flash Attention 2
|
|||||||
|
|
||||||
`Flash Attention <https://github.com/Dao-AILab/flash-attention>`_ supports two backend implementations on AMD GPUs.
|
`Flash Attention <https://github.com/Dao-AILab/flash-attention>`_ supports two backend implementations on AMD GPUs.
|
||||||
|
|
||||||
* `Composable Kernel (CK) <https://github.com/ROCm/composable_kernel>`_ - the default backend
|
* `Composable Kernel (CK) <https://github.com/ROCm/composable_kernel>`__ - the default backend
|
||||||
* `OpenAI Triton <https://github.com/triton-lang/triton>`_ - an alternative backend
|
* `OpenAI Triton <https://github.com/triton-lang/triton>`__ - an alternative backend
|
||||||
|
|
||||||
You can switch between these backends using the environment variable FLASH_ATTENTION_TRITON_AMD_ENABLE:
|
You can switch between these backends using the environment variable ``FLASH_ATTENTION_TRITON_AMD_ENABLE``:
|
||||||
|
|
||||||
FLASH_ATTENTION_TRITON_AMD_ENABLE="FALSE"
|
``FLASH_ATTENTION_TRITON_AMD_ENABLE="FALSE"``
|
||||||
→ Use Composable Kernel (CK) backend (FlashAttention 2)
|
→ Use Composable Kernel (CK) backend (Flash Attention 2)
|
||||||
|
|
||||||
FLASH_ATTENTION_TRITON_AMD_ENABLE="TRUE"
|
``FLASH_ATTENTION_TRITON_AMD_ENABLE="TRUE"``
|
||||||
→ Use OpenAI Triton backend (FlashAttention 2)
|
→ Use OpenAI Triton backend (Flash Attention 2)
|
||||||
|
|
||||||
To install Flash Attention 2, use the following commands:
|
To install Flash Attention 2, use the following commands:
|
||||||
|
|
||||||
@@ -74,7 +74,7 @@ For detailed installation instructions, see `Flash Attention <https://github.com
|
|||||||
Benchmarking Flash Attention 2
|
Benchmarking Flash Attention 2
|
||||||
------------------------------
|
------------------------------
|
||||||
|
|
||||||
Benchmark scripts to evaluate the performance of Flash Attention 2 are stored in the `flash-attention/benchmarks/` directory.
|
Benchmark scripts to evaluate the performance of Flash Attention 2 are stored in the ``flash-attention/benchmarks/`` directory.
|
||||||
|
|
||||||
To benchmark the CK backend
|
To benchmark the CK backend
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user