Files
tinygrad/extra
Yixiang Gao 13e872b53f add mutigpu support for llama attention (#3064)
* add llama attention test for multigpu

* test fails

* kv cache trying to shrink on sharded axis

* mask None works for scale dot product

* kv cache seems to be working but scale dot product breaks

* scaled dot product works, but the last linear layer failed

* running into the reshape case where it could be wrong for multigpu

* making sure it was the reshape

* adding contiguous doesn't solve

* need to shard more properly

* remove reshape test

* minor adjustment to scale dot product attention test

* weights are sharded wrong

* continue fix new weight sharding

* clean up

* fix attention when start_pos is 0

* remove print

* add TODOs for the best mutigpu interface
2024-01-11 16:31:02 -08:00
..
2023-11-13 20:18:40 -08:00
2024-01-01 14:58:48 -08:00
2024-01-08 09:29:13 -08:00
2023-12-01 09:25:27 -08:00
2024-01-04 13:59:43 -08:00
2024-01-01 14:58:48 -08:00
2023-12-25 22:52:49 -05:00
2024-01-05 10:33:13 -08:00
2024-01-08 09:29:13 -08:00
2023-07-05 15:36:22 -07:00
2023-12-07 17:07:05 -08:00
2024-01-04 02:02:42 -05:00
2023-11-30 17:07:16 -08:00
2024-01-09 19:28:09 -08:00