mirror of
https://github.com/ROCm/ROCm.git
synced 2026-02-12 15:34:58 -05:00
[docs/7.11.0] Add llama.cpp known issue (#5962)
This commit is contained in:
13
RELEASE.md
13
RELEASE.md
@@ -1729,6 +1729,19 @@ known to be affected. As a workaround, use `-Og` optimization level instead of
|
||||
llama.cpp builds successfully but might fail at runtime with the error "HIP
|
||||
kernel mul_mat_q has no device code". This issue is under investigation.
|
||||
|
||||
### llama.cpp prompt processing performance regression
|
||||
|
||||
llama.cpp experiences reduced prompt processing performance across multiple AMD
|
||||
GPU architectures. As a workaround, pass the compiler flag `-mllvm
|
||||
--amdgpu-unroll-threshold-local=600` to `hipcc` or `amdclang` when compiling
|
||||
llama.cpp:
|
||||
|
||||
```bash
|
||||
hipcc -mllvm --amdgpu-unroll-threshold-local=600 ...
|
||||
```
|
||||
|
||||
This issue will be fixed in a future release.
|
||||
|
||||
### PyTorch model training validation issues
|
||||
|
||||
The following models failed validation on PyTorch for ROCm 7.11.0 due to
|
||||
|
||||
Reference in New Issue
Block a user