George Hotz
088f280dc3
touchups
2020-12-07 07:50:27 -08:00
George Hotz
0cf21881b7
hwx parse w/o macho mods
2020-12-06 23:13:28 -08:00
Josh Smith
aa4161f63e
use classmethods for Tensor helper funcs ( #146 )
2020-12-06 22:35:43 -08:00
George Hotz
e75a6d1b4b
quadconv
2020-12-06 20:39:50 -08:00
George Hotz
23664c99bd
double conv
2020-12-06 20:26:02 -08:00
George Hotz
1a0f826dc6
highlight the commands
2020-12-06 20:03:21 -08:00
George Hotz
2f1f006003
we have docs
2020-12-06 19:54:03 -08:00
George Hotz
3531e81f0e
dumping ANE docs
2020-12-06 18:58:36 -08:00
George Hotz
6e793e96c3
deeebug
2020-12-06 17:49:17 -08:00
George Hotz
dced0cb44b
oops, path to weights
2020-12-06 16:33:42 -08:00
George Hotz
c57dc61ea7
simple op examples
2020-12-06 16:32:26 -08:00
George Hotz
7babf38617
found concat
2020-12-06 16:27:12 -08:00
George Hotz
ddd6778423
add neuron
2020-12-06 16:24:42 -08:00
George Hotz
f2f2d6aea3
docs and noop
2020-12-06 16:10:44 -08:00
George Hotz
6ba25834ee
found some plists
2020-12-06 15:57:28 -08:00
George Hotz
e2184c20ad
min weights, update golden
2020-12-06 15:29:15 -08:00
George Hotz
d4d8bd0337
make minimal plist for compare
2020-12-06 15:10:15 -08:00
George Hotz
0845ec43c6
compile takes in plist
2020-12-06 14:51:33 -08:00
George Hotz
00312b8ad1
batchnorm work
2020-12-06 14:40:07 -08:00
George Hotz
da514c2918
fix enet init
2020-12-06 13:52:07 -08:00
George Hotz
3b982f2f7a
get_parameters
2020-12-06 13:47:28 -08:00
George Hotz
102e6356e9
replace layer_init_uniform with .uniform
2020-12-06 13:44:31 -08:00
George Hotz
c71a8ef222
remove unused Div op
2020-12-06 13:02:12 -08:00
George Hotz
20f95de408
less lines
2020-12-06 12:56:03 -08:00
George Hotz
629efb391f
lose a few lines
2020-12-06 12:42:41 -08:00
George Hotz
521098cc2f
se optional, track time better
2020-12-06 12:29:42 -08:00
George Hotz
609d11e699
trainer works with CIFAR
2020-12-06 12:20:14 -08:00
George Hotz
80a9c777ba
requires grad, optim in train enet
2020-12-06 11:10:30 -08:00
George Hotz
c66c27d22e
get parameters
2020-12-06 10:45:04 -08:00
George Hotz
51daaa43d4
fix memory leaks, add gc test
2020-12-06 10:34:40 -08:00
George Hotz
1717daa859
reshape doesn't copy anymore
2020-12-06 09:51:09 -08:00
George Hotz
62ee47fef8
add GPUBuffer class
2020-12-06 09:45:13 -08:00
George Hotz
17659f7dd7
gpu speedup, tests work on M1
2020-12-06 09:05:49 -08:00
George Hotz
b8deb36e56
train BS=16 for 32 steps
2020-12-04 10:00:32 -08:00
George Hotz
ad1b225722
oops, i broke fill
2020-12-04 09:53:38 -08:00
George Hotz
cb79c9838e
make the GPU 25% faster by not recreating kernels
2020-12-04 09:51:00 -08:00
George Hotz
df81bf5985
it's the default, but make it explicit
2020-12-04 09:43:41 -08:00
George Hotz
888689b57b
proprotip
2020-12-04 09:24:46 -08:00
George Hotz
2862b42bac
install from github
2020-12-04 09:06:25 -08:00
George Hotz
1290e01e2c
all ops supported on GPU now
2020-12-03 10:43:11 -08:00
George Hotz
621a93b777
ane in readme
2020-12-03 10:40:31 -08:00
George Hotz
1dcaecacc4
Support for Apple Neural Engine ( #130 )
...
* ane query is success
* cite and build instructions
* low level access, need to disable AMFI
* coreml_ane works
* coreml fun
* more work
* compiled example
* progress
* compiler works
* model flow
* TODOs in the readme
* put some real weights in
* we are learning objc
* much progress i think
* signed model still doesn't work
* working example
* there are float16
* clean up: part 1
* h11ane header, more cleanup
* cleanup DeviceController creation
* remove the stupid sleep
* notes
* start a hwx parser
* no tabs
* compare stuff
* hmm, why don't inputs work
* cache doesn't seem to fix it
* hmm, the issue was the compiler
* fix the compiler, guess i didn't put in weights
* logging for compiler
* uselessness in plist
* remove hwx before compile, weights are converted to float16
* better compare
* better compare
* last line in comparE
* opcodes from compiler
* notes
2020-12-03 10:32:26 -08:00
baplou
c83cebccda
Made the readme more consistent ( #136 )
2020-11-28 08:20:02 -06:00
Marcel Bischoff
541330c42a
Update README.md ( #133 )
...
should we put `ipython3` otherwise the path doesn't work or we have to add the env, not sure what is nicer
2020-11-25 07:53:54 -08:00
Mufeed VH
0bbf66627c
Define ProfileOp class once ( #131 )
...
* define `ProfileOp` class once
* clean `ProfileOp` class
* removed `else: pass`
2020-11-24 19:39:13 -08:00
George Hotz
03994e0011
load torch files without torch
2020-11-21 13:43:53 -08:00
Marcel Bischoff
26899869a2
Update tensor.py ( #128 )
...
Otherwise `.cpu()` is broken if default is GPU
2020-11-21 09:16:03 -08:00
adamritter
f190ca446d
Detach ( #123 )
...
* Detach
* Torch.detach reuses the buffer in the
* Fix test
* wakey wakey GitHub Actions
Co-authored-by: holonomicjl <58403584+holonomicjl@users.noreply.github.com >
2020-11-19 19:03:42 -08:00
Colin Manko
8383ff40ad
fix pyopencl ( #125 )
2020-11-19 19:03:04 -08:00
adamritter
5797e63d9b
Train efficientnet should respect NUM environment variable ( #122 )
...
Co-authored-by: holonomicjl <58403584+holonomicjl@users.noreply.github.com >
2020-11-16 20:02:31 -08:00