George Hotz
dca076dbf1
remove dumb nn ops
2021-11-29 18:05:31 -05:00
George Hotz
33720e733f
support keepdim
2021-11-29 17:47:43 -05:00
George Hotz
8a02bd56a1
refactor: canonicalize axis
2021-11-29 17:29:18 -05:00
George Hotz
70544e7e9f
sum hook override
2021-11-29 17:14:24 -05:00
George Hotz
8097b8f7d6
vit works
2021-11-29 16:28:14 -05:00
George Hotz
7c07c5efdd
plz fix vit
2021-11-29 15:45:19 -05:00
George Hotz
ca160504e1
affine is always the last dim
2021-11-29 15:22:49 -05:00
George Hotz
e86f7a4aa3
deterministic
2021-11-29 15:10:15 -05:00
George Hotz
f909ab194f
gelu with broken test
2021-11-29 15:00:50 -05:00
George Hotz
9ce881f88c
fix bug in getitem, drop int axis
2021-11-29 14:01:24 -05:00
George Hotz
c752033283
fix GPU OOM in test
2021-11-29 13:05:59 -05:00
George Hotz
1eafa5580e
layernorm with learnable parameters
2021-11-29 13:03:57 -05:00
George Hotz
c7f795ca1e
added dot affine
2021-11-29 12:55:56 -05:00
George Hotz
30eb3afbe1
add bias term to transformer
2021-11-29 12:45:27 -05:00
George Hotz
99b6051467
add ff_dim to transformer
2021-11-29 12:40:52 -05:00
George Hotz
b4839eb6bb
factor out binary_broadcast
2021-11-29 03:49:12 -05:00
George Hotz
ca154f8b59
add sub and a few refactors
2021-11-29 03:35:23 -05:00
George Hotz
7b3c442f3b
sum in metal
2021-11-29 03:13:41 -05:00
George Hotz
c6fb087695
fix old list behavior
2021-11-29 01:49:21 -05:00
George Hotz
7c160c6cee
metal: got relu working and add somewhat working
2021-11-29 01:40:18 -05:00
George Hotz
29dee59368
cat: forward only not required
2021-11-29 00:14:56 -05:00
George Hotz
06838481e8
use get data dtype
2021-11-28 23:28:37 -05:00
George Hotz
3cdc77f526
add cat support
2021-11-28 23:21:49 -05:00
George Hotz
ce3d198bb7
less lines and fix default device
2021-11-27 11:18:49 -05:00
George Hotz
7ae14179d3
refactor ops
2021-11-27 11:12:23 -05:00
George Hotz
4320c45c4b
oops, gpu not available
2021-11-17 12:54:33 -08:00
George Hotz
e2a8961a18
less lines, fix bug
2021-11-17 12:52:17 -08:00
George Hotz
e8ca3ad053
add Buffer abstraction for each device
2021-11-17 12:46:07 -08:00
George Hotz
c162e748f5
fix float64 warning on training
2021-10-30 20:07:31 -07:00
George Hotz
b0f14b4af8
move datasets into datasets
2021-10-30 19:55:50 -07:00
George Hotz
a95ef16c8c
sub 1000 lines
2021-10-30 19:48:24 -07:00
George Hotz
844540a5ed
yolo in readme
2021-10-30 19:47:34 -07:00
George Hotz
ba28761894
move yolo into examples/yolo
2021-10-30 19:46:00 -07:00
George Hotz
121d5a17ee
use tinynn for Conv2d
2021-10-30 19:40:44 -07:00
George Hotz
6bee5bdb7d
add torch tests
2021-10-30 18:58:45 -07:00
George Hotz
75260c6e49
unify more of cpu and torch
2021-10-30 18:57:24 -07:00
George Hotz
b1dae764d3
more similar between torch and numpy
2021-10-30 18:46:59 -07:00
George Hotz
d6cb373c79
import the torch ops that are the same
2021-10-30 18:22:44 -07:00
George Hotz
51a49de299
tests pass for torch ops
2021-10-30 18:00:12 -07:00
George Hotz
641b1dbb40
remove ane, start supporting ops_torch
2021-10-30 17:47:00 -07:00
George Hotz
7d12482d80
refactor efficientnet loading
2021-10-30 17:02:17 -07:00
George Hotz
7472a7ebe2
not forcing 3.9 for a stupid type
2021-10-30 16:52:40 -07:00
George Hotz
114f6ca3fd
more readme cleanup
2021-10-30 16:51:25 -07:00
George Hotz
fc6597a6d9
only resnet18, it's too slow otherwise
2021-10-30 16:48:39 -07:00
George Hotz
c05867dcbb
header
2021-10-30 16:41:05 -07:00
George Hotz
e10dc90524
tpu docs
2021-10-30 16:38:19 -07:00
George Hotz
effd0dc833
update readme
2021-10-30 16:34:00 -07:00
George Hotz
f193eeed25
bump all to python 3.9
2021-10-30 16:15:41 -07:00
George Hotz
63f50cff45
move back again
2021-10-30 16:13:29 -07:00
Evan Mays
285621aeda
Cherry backprop for conv2d ( #281 )
...
* quick math: 0 + x = x.
* gradient w.r.t. x using cherry for conv
* gradient w.r.t. w for conv on cherry but doing vector dot products
* small optimization
* [cherry] optimize conv backpass for large channel count
* get rid of numpy einsum
2021-10-30 16:12:19 -07:00