[FRONTEND] Add input dtypes to autotuning key (#2534) (#374)

* [FRONTEND] Add input dtypes to autotuning key (#2534)

* Fix conflict in 06-fused-attention

* Fix get_best_config in FA-transV.py

* Fix leftover get_best_config()

---------

Co-authored-by: Adnan Akhundov <adnan.akhundov@gmail.com>
This commit is contained in:
Lixun Zhang
2023-11-07 19:36:57 -06:00
committed by GitHub
parent 3c1fe617c1
commit 1af893d8a2
5 changed files with 10 additions and 20 deletions

View File

@@ -236,10 +236,6 @@ def matmul(a, b, activation=""):
return c
def get_best_config(M, N, K):
best_config = matmul_kernel_splitK.get_best_config(M = M, N = N, K = K)
return best_config
def test_correctness(M, N, K, datatype = torch.float16):
torch.manual_seed(0)
@@ -341,7 +337,7 @@ def main():
if args.compare:
test_correctness(m, n, k, dtype)
best_config = get_best_config(m, n, k)
best_config = matmul_kernel_splitK.get_best_config()
if use_rocprof:
dtype_str = 'fp16' if (not args.specify_type) else args.dtype