Files
tinygrad/test
chenyu 4a6d84c4c3 hotfix llama start_pos vmax is max_context-1 (#10659)
* hotfix llama start_pos vmax is max_context-1

fixed `IGNORE_OOB=0 python3 examples/llama3.py --size 1B --benchmark --temperature 0`

* hotfix: multitensor transformer test tests kv cache

---------

Co-authored-by: George Hotz <geohot@gmail.com>
2025-06-06 00:41:25 -04:00
..
2025-06-05 17:17:42 -07:00
2025-02-20 18:03:09 -05:00
2025-05-26 14:38:28 -07:00
2025-06-04 14:58:24 +03:00
2020-12-15 23:44:08 -08:00
2024-11-11 20:18:04 +08:00
2025-02-17 14:47:54 +01:00
2025-02-20 18:03:09 -05:00
2025-02-18 15:26:58 +08:00
2025-04-22 19:12:49 -04:00
2025-02-20 18:03:09 -05:00
2025-05-18 11:38:28 -07:00
2025-05-07 11:41:41 -07:00
2025-05-07 11:41:41 -07:00
2025-02-20 18:03:09 -05:00
2023-12-07 17:07:05 -08:00
2025-05-18 11:38:28 -07:00
2025-06-05 17:17:42 -07:00
2025-02-20 18:03:09 -05:00