George Hotz
b132de677d
tinygrad.nn ( #367 )
...
* tinygrad.nn
* flake8
* working on pylint
* more pylint
* more pylint
* pylint passes
* networkx
* mypy can't infer that type
* junk
2022-08-18 07:41:00 -07:00
George Hotz
acbeaf0ba9
adam in benchmark_train_efficientnet
2022-07-19 09:33:07 -07:00
George Hotz
d985217fa4
skip reduce noops
2022-07-16 07:47:43 -07:00
George Hotz
5e46561f7e
no_grad = NOT backward
2022-07-10 20:54:57 -07:00
George Hotz
d5d9cffe7c
training param for batchnorm
2022-07-04 13:28:03 -07:00
George Hotz
34f43ea10e
LAZY and CLCACHE are defaults
2022-07-04 13:09:15 -07:00
George Hotz
b7afd83267
track cl mem used
2022-07-04 12:19:00 -07:00
George Hotz
d5de8452c6
dashed loadops
2022-07-04 09:50:56 -07:00
George Hotz
7276f8d6bf
improve constant folding, detach before moving tensor
2022-07-02 15:29:40 -07:00
George Hotz
0cb99d72e9
NUM=-1 is a small efficientnet for small people
2022-07-02 15:11:51 -07:00
George Hotz
8cf1aed0f4
don't track_running_stats, parameters must require_grad
2022-07-02 14:38:45 -07:00
George Hotz
f607f18006
fix backward
2022-06-25 00:00:53 -07:00
George Hotz
ec30f0402f
improve benchmark_train_efficientnet
2022-06-24 23:46:38 -07:00
George Hotz
d748353ce5
err, okay, a bit more off
2022-06-24 22:44:57 -07:00
George Hotz
bdde95f16e
CACHE_LAZYBUFFERS options + benchmark. only a couple x from torch
2022-06-24 22:33:53 -07:00
Edward Betts
a6203b5ecd
Correct spelling mistakes ( #314 )
2022-04-05 05:22:18 -07:00
Ildar Kashaev
d065040df0
Fix yolo param name ( #306 )
2022-02-10 01:40:24 -07:00
George Hotz
2a10116bfa
support drawing graphs
2022-01-16 10:45:58 -08:00
George Hotz
2cae2dfa07
don't crash the dataloader for imagenet
2022-01-16 08:41:26 -08:00
George Hotz
907ff7dbb6
imagenet is training
2022-01-15 23:27:31 -08:00
George Hotz
d1e082e0ef
add imagenet training support
2022-01-15 23:16:38 -08:00
George Hotz
ade2af7ac0
data doesn't require grad
2022-01-15 22:41:27 -08:00
George Hotz
0973e54eb0
fix DEBUG for GPU
2022-01-15 22:14:28 -08:00
cjg91
7025c9bbeb
Transfer learning for ResNet ( #295 )
...
* Transfer learning for ResNet
* moved ResNet depth specifics into the class
2022-01-15 23:22:10 -05:00
George Hotz
55d792b065
Revert "fixup resnet"
...
This reverts commit 4eabe677ed .
2022-01-15 20:22:01 -08:00
George Hotz
4eabe677ed
fixup resnet
2022-01-15 20:21:02 -08:00
George Hotz
e0bef0bd01
training is False by default
2022-01-15 19:57:41 -08:00
George Hotz
8ec2341cca
fix bn training
2022-01-15 19:47:01 -08:00
George Hotz
c0c2c0b041
support larger ViT models
2021-12-12 10:45:10 -08:00
George Hotz
4b7c31b5b7
break vit into it's own file
2021-11-30 11:19:22 -05:00
George Hotz
46bbbcf7f0
model touchups
2021-11-30 11:13:34 -05:00
George Hotz
835869974c
clean up vit code
2021-11-30 10:58:03 -05:00
George Hotz
535f02cc64
use sequential
2021-11-30 00:25:39 -05:00
George Hotz
de938c2d9d
vit is now tested
2021-11-30 00:23:06 -05:00
George Hotz
aff810e722
unify transformer block
2021-11-29 18:58:15 -05:00
George Hotz
58ed46963e
fix broadcastdot
2021-11-29 18:54:57 -05:00
George Hotz
125e74293f
promote layernorm to tensor op
2021-11-29 18:08:21 -05:00
George Hotz
dca076dbf1
remove dumb nn ops
2021-11-29 18:05:31 -05:00
George Hotz
70544e7e9f
sum hook override
2021-11-29 17:14:24 -05:00
George Hotz
8097b8f7d6
vit works
2021-11-29 16:28:14 -05:00
George Hotz
7c07c5efdd
plz fix vit
2021-11-29 15:45:19 -05:00
George Hotz
ca160504e1
affine is always the last dim
2021-11-29 15:22:49 -05:00
George Hotz
e86f7a4aa3
deterministic
2021-11-29 15:10:15 -05:00
George Hotz
f909ab194f
gelu with broken test
2021-11-29 15:00:50 -05:00
George Hotz
9ce881f88c
fix bug in getitem, drop int axis
2021-11-29 14:01:24 -05:00
George Hotz
99b6051467
add ff_dim to transformer
2021-11-29 12:40:52 -05:00
George Hotz
b0f14b4af8
move datasets into datasets
2021-10-30 19:55:50 -07:00
George Hotz
ba28761894
move yolo into examples/yolo
2021-10-30 19:46:00 -07:00
George Hotz
121d5a17ee
use tinynn for Conv2d
2021-10-30 19:40:44 -07:00
Guglielmo Camporese
2b7589db64
Added ResNet-{18, 34, 50, 101, 152} ( #271 )
...
* added resnets
* fix minor
* fix minor
* resnet in models
* added resnet test
* added resnet train test
* added linear, conv2d nn tests
* fix minor in extra/training
* resnet in models
* fix minor
* fix tolerance for linear in nn test
* fix eval, this causes cpu and gpu UT failing
* revert transformer test
* fix minor for CPU test
* improved model get_params for sequential layer
* fix minor for params counting
* commented broken ops tests
* improved train for resnet
2021-06-21 09:37:24 -07:00