Commit Graph

11106 Commits

Author SHA1 Message Date
George Hotz
088f280dc3 touchups 2020-12-07 07:50:27 -08:00
George Hotz
0cf21881b7 hwx parse w/o macho mods 2020-12-06 23:13:28 -08:00
Josh Smith
aa4161f63e use classmethods for Tensor helper funcs (#146) 2020-12-06 22:35:43 -08:00
George Hotz
e75a6d1b4b quadconv 2020-12-06 20:39:50 -08:00
George Hotz
23664c99bd double conv 2020-12-06 20:26:02 -08:00
George Hotz
1a0f826dc6 highlight the commands 2020-12-06 20:03:21 -08:00
George Hotz
2f1f006003 we have docs 2020-12-06 19:54:03 -08:00
George Hotz
3531e81f0e dumping ANE docs 2020-12-06 18:58:36 -08:00
George Hotz
6e793e96c3 deeebug 2020-12-06 17:49:17 -08:00
George Hotz
dced0cb44b oops, path to weights 2020-12-06 16:33:42 -08:00
George Hotz
c57dc61ea7 simple op examples 2020-12-06 16:32:26 -08:00
George Hotz
7babf38617 found concat 2020-12-06 16:27:12 -08:00
George Hotz
ddd6778423 add neuron 2020-12-06 16:24:42 -08:00
George Hotz
f2f2d6aea3 docs and noop 2020-12-06 16:10:44 -08:00
George Hotz
6ba25834ee found some plists 2020-12-06 15:57:28 -08:00
George Hotz
e2184c20ad min weights, update golden 2020-12-06 15:29:15 -08:00
George Hotz
d4d8bd0337 make minimal plist for compare 2020-12-06 15:10:15 -08:00
George Hotz
0845ec43c6 compile takes in plist 2020-12-06 14:51:33 -08:00
George Hotz
00312b8ad1 batchnorm work 2020-12-06 14:40:07 -08:00
George Hotz
da514c2918 fix enet init 2020-12-06 13:52:07 -08:00
George Hotz
3b982f2f7a get_parameters 2020-12-06 13:47:28 -08:00
George Hotz
102e6356e9 replace layer_init_uniform with .uniform 2020-12-06 13:44:31 -08:00
George Hotz
c71a8ef222 remove unused Div op 2020-12-06 13:02:12 -08:00
George Hotz
20f95de408 less lines 2020-12-06 12:56:03 -08:00
George Hotz
629efb391f lose a few lines 2020-12-06 12:42:41 -08:00
George Hotz
521098cc2f se optional, track time better 2020-12-06 12:29:42 -08:00
George Hotz
609d11e699 trainer works with CIFAR 2020-12-06 12:20:14 -08:00
George Hotz
80a9c777ba requires grad, optim in train enet 2020-12-06 11:10:30 -08:00
George Hotz
c66c27d22e get parameters 2020-12-06 10:45:04 -08:00
George Hotz
51daaa43d4 fix memory leaks, add gc test 2020-12-06 10:34:40 -08:00
George Hotz
1717daa859 reshape doesn't copy anymore 2020-12-06 09:51:09 -08:00
George Hotz
62ee47fef8 add GPUBuffer class 2020-12-06 09:45:13 -08:00
George Hotz
17659f7dd7 gpu speedup, tests work on M1 2020-12-06 09:05:49 -08:00
George Hotz
b8deb36e56 train BS=16 for 32 steps 2020-12-04 10:00:32 -08:00
George Hotz
ad1b225722 oops, i broke fill 2020-12-04 09:53:38 -08:00
George Hotz
cb79c9838e make the GPU 25% faster by not recreating kernels 2020-12-04 09:51:00 -08:00
George Hotz
df81bf5985 it's the default, but make it explicit 2020-12-04 09:43:41 -08:00
George Hotz
888689b57b proprotip 2020-12-04 09:24:46 -08:00
George Hotz
2862b42bac install from github 2020-12-04 09:06:25 -08:00
George Hotz
1290e01e2c all ops supported on GPU now 2020-12-03 10:43:11 -08:00
George Hotz
621a93b777 ane in readme 2020-12-03 10:40:31 -08:00
George Hotz
1dcaecacc4 Support for Apple Neural Engine (#130)
* ane query is success

* cite and build instructions

* low level access, need to disable AMFI

* coreml_ane works

* coreml fun

* more work

* compiled example

* progress

* compiler works

* model flow

* TODOs in the readme

* put some real weights in

* we are learning objc

* much progress i think

* signed model still doesn't work

* working example

* there are float16

* clean up: part 1

* h11ane header, more cleanup

* cleanup DeviceController creation

* remove the stupid sleep

* notes

* start a hwx parser

* no tabs

* compare stuff

* hmm, why don't inputs work

* cache doesn't seem to fix it

* hmm, the issue was the compiler

* fix the compiler, guess i didn't put in weights

* logging for compiler

* uselessness in plist

* remove hwx before compile, weights are converted to float16

* better compare

* better compare

* last line in comparE

* opcodes from compiler

* notes
2020-12-03 10:32:26 -08:00
baplou
c83cebccda Made the readme more consistent (#136) 2020-11-28 08:20:02 -06:00
Marcel Bischoff
541330c42a Update README.md (#133)
should we put `ipython3` otherwise the path doesn't work or we have to add the env, not sure what is nicer
2020-11-25 07:53:54 -08:00
Mufeed VH
0bbf66627c Define ProfileOp class once (#131)
* define `ProfileOp` class once

* clean `ProfileOp` class

* removed `else: pass`
2020-11-24 19:39:13 -08:00
George Hotz
03994e0011 load torch files without torch 2020-11-21 13:43:53 -08:00
Marcel Bischoff
26899869a2 Update tensor.py (#128)
Otherwise `.cpu()` is broken if default is GPU
2020-11-21 09:16:03 -08:00
adamritter
f190ca446d Detach (#123)
* Detach

* Torch.detach reuses the buffer in the

* Fix test

* wakey wakey GitHub Actions

Co-authored-by: holonomicjl <58403584+holonomicjl@users.noreply.github.com>
2020-11-19 19:03:42 -08:00
Colin Manko
8383ff40ad fix pyopencl (#125) 2020-11-19 19:03:04 -08:00
adamritter
5797e63d9b Train efficientnet should respect NUM environment variable (#122)
Co-authored-by: holonomicjl <58403584+holonomicjl@users.noreply.github.com>
2020-11-16 20:02:31 -08:00