Files
tinygrad/extra/models
chenyu 4a6d84c4c3 hotfix llama start_pos vmax is max_context-1 (#10659)
* hotfix llama start_pos vmax is max_context-1

fixed `IGNORE_OOB=0 python3 examples/llama3.py --size 1B --benchmark --temperature 0`

* hotfix: multitensor transformer test tests kv cache

---------

Co-authored-by: George Hotz <geohot@gmail.com>
2025-06-06 00:41:25 -04:00
..
2025-04-03 05:44:09 -04:00
2024-10-07 00:47:32 +08:00
2024-10-01 19:47:58 -04:00
2024-05-24 17:04:19 -04:00
2024-09-24 10:08:04 +08:00
2023-11-28 17:36:55 -08:00
2023-11-28 17:36:55 -08:00