Commit Graph

1207 Commits

Author SHA1 Message Date
George Hotz
9590d92750 stable diffusion compiles (add no_init) 2022-09-04 11:40:50 -07:00
George Hotz
172683c314 work 2022-09-04 11:21:09 -07:00
George Hotz
c2a030fe55 one liner that's more clear 2022-09-03 16:08:48 -07:00
George Hotz
4a3ed58edb more readable actually 2022-09-03 16:00:35 -07:00
George Hotz
633f31dc73 easier to read 2022-09-03 15:53:58 -07:00
George Hotz
6578e08919 cleanups for Mid 2022-09-03 15:50:33 -07:00
George Hotz
852de7c66c remove ugly parens 2022-09-03 15:41:37 -07:00
George Hotz
6b190c2fa5 stable diffusion works 2022-09-03 13:55:36 -07:00
George Hotz
947e10dab0 yolo 2022-09-03 12:39:48 -07:00
George Hotz
033a3ecccf found tinygrad bug 2022-09-03 12:32:43 -07:00
George Hotz
114728d363 torch bs 2022-09-03 11:57:23 -07:00
George Hotz
356732515b stable_diffusion: add attn and layernorm 2022-09-03 11:02:27 -07:00
George Hotz
4dadd95e3c fix tests hopefully, more stable diffusion 2022-09-03 10:38:31 -07:00
George Hotz
c01a8c5c2d stable diffusion start 2022-09-03 10:08:42 -07:00
George Hotz
b132de677d tinygrad.nn (#367)
* tinygrad.nn

* flake8

* working on pylint

* more pylint

* more pylint

* pylint passes

* networkx

* mypy can't infer that type

* junk
2022-08-18 07:41:00 -07:00
George Hotz
acbeaf0ba9 adam in benchmark_train_efficientnet 2022-07-19 09:33:07 -07:00
George Hotz
d985217fa4 skip reduce noops 2022-07-16 07:47:43 -07:00
George Hotz
5e46561f7e no_grad = NOT backward 2022-07-10 20:54:57 -07:00
George Hotz
d5d9cffe7c training param for batchnorm 2022-07-04 13:28:03 -07:00
George Hotz
34f43ea10e LAZY and CLCACHE are defaults 2022-07-04 13:09:15 -07:00
George Hotz
b7afd83267 track cl mem used 2022-07-04 12:19:00 -07:00
George Hotz
d5de8452c6 dashed loadops 2022-07-04 09:50:56 -07:00
George Hotz
7276f8d6bf improve constant folding, detach before moving tensor 2022-07-02 15:29:40 -07:00
George Hotz
0cb99d72e9 NUM=-1 is a small efficientnet for small people 2022-07-02 15:11:51 -07:00
George Hotz
8cf1aed0f4 don't track_running_stats, parameters must require_grad 2022-07-02 14:38:45 -07:00
George Hotz
f607f18006 fix backward 2022-06-25 00:00:53 -07:00
George Hotz
ec30f0402f improve benchmark_train_efficientnet 2022-06-24 23:46:38 -07:00
George Hotz
d748353ce5 err, okay, a bit more off 2022-06-24 22:44:57 -07:00
George Hotz
bdde95f16e CACHE_LAZYBUFFERS options + benchmark. only a couple x from torch 2022-06-24 22:33:53 -07:00
Edward Betts
a6203b5ecd Correct spelling mistakes (#314) 2022-04-05 05:22:18 -07:00
Ildar Kashaev
d065040df0 Fix yolo param name (#306) 2022-02-10 01:40:24 -07:00
George Hotz
2a10116bfa support drawing graphs 2022-01-16 10:45:58 -08:00
George Hotz
2cae2dfa07 don't crash the dataloader for imagenet 2022-01-16 08:41:26 -08:00
George Hotz
907ff7dbb6 imagenet is training 2022-01-15 23:27:31 -08:00
George Hotz
d1e082e0ef add imagenet training support 2022-01-15 23:16:38 -08:00
George Hotz
ade2af7ac0 data doesn't require grad 2022-01-15 22:41:27 -08:00
George Hotz
0973e54eb0 fix DEBUG for GPU 2022-01-15 22:14:28 -08:00
cjg91
7025c9bbeb Transfer learning for ResNet (#295)
* Transfer learning for ResNet

* moved ResNet depth specifics into the class
2022-01-15 23:22:10 -05:00
George Hotz
55d792b065 Revert "fixup resnet"
This reverts commit 4eabe677ed.
2022-01-15 20:22:01 -08:00
George Hotz
4eabe677ed fixup resnet 2022-01-15 20:21:02 -08:00
George Hotz
e0bef0bd01 training is False by default 2022-01-15 19:57:41 -08:00
George Hotz
8ec2341cca fix bn training 2022-01-15 19:47:01 -08:00
George Hotz
c0c2c0b041 support larger ViT models 2021-12-12 10:45:10 -08:00
George Hotz
4b7c31b5b7 break vit into it's own file 2021-11-30 11:19:22 -05:00
George Hotz
46bbbcf7f0 model touchups 2021-11-30 11:13:34 -05:00
George Hotz
835869974c clean up vit code 2021-11-30 10:58:03 -05:00
George Hotz
535f02cc64 use sequential 2021-11-30 00:25:39 -05:00
George Hotz
de938c2d9d vit is now tested 2021-11-30 00:23:06 -05:00
George Hotz
aff810e722 unify transformer block 2021-11-29 18:58:15 -05:00
George Hotz
58ed46963e fix broadcastdot 2021-11-29 18:54:57 -05:00