Commit Graph

7979 Commits

Author SHA1 Message Date
George Hotz
ffb96b2d0b batchnorm by marcelbischoff 2020-12-09 03:23:04 -08:00
NeuralLink
00e376f36c leaky relu as geohot suggested (#167) 2020-12-09 02:58:35 -08:00
George Hotz
c225e62dd2 touchups 2020-12-09 02:52:28 -08:00
Liam
89d0ff6989 Consistent testing (#137)
* Consistent GPU classes

Convert the existing GPU classes into one standard format.

Remove duplicated functions in `test_mnist` and create a TestMNISTGPU
class. This reduces line count and ensures consistency.

Use `@unittest.skipUnless(GPU, "Requires GPU")` instead of `if GPU:` to
skip GPU testing. This will ensure that skipped tests are displayed
accordingly in the pytest output.

* Optim Testing now supports GPU

* Tensor testing now supports GPU

jacobian and gradcheck auto skipped until GPU float64 support added.

* GPU support for custom constructor methods

* Remove GPU flag from Model constructors

It was requested that the `gpu` kwarg be removed from the model
constructor. GPU conversion is now handled in the train function.

This also required the conversion of Optimizer parameters as they are
constructed prior to execution of the `train` function and are dependant
on the model GPU state.

* Fix typo: float32->float64

* Clean `get_parameters` utility

Just a quick refactor w/ the new support for optimizers.

* Remove GPU kwarg from TinyNet

Remove `gpu` kwarg from tiny net to match test_mnist `train` function.
2020-12-09 02:25:27 -08:00
Liam
34b38dd4d0 Extra install requirements. (#164)
* Testing install requirements

* GPU install requirements
2020-12-09 02:22:47 -08:00
George Hotz
0e02f394ee serious_mnist 2020-12-08 21:43:05 -08:00
Daulet
24d688c184 win more lines for core library (#158)
...and sacrifice test speed
2020-12-08 14:18:45 -08:00
NeuralLink
9f77fd6135 🔨 refactor optim (#156)
* 🔨 refactor optim

* 🔨 refactor optim

* 🔨 more clean up
2020-12-08 14:16:31 -08:00
George Hotz
4e1a0de392 fix rsub 2020-12-08 10:05:21 -08:00
George Hotz
c4540f1b8c Support scalars by kartik4949 2020-12-08 09:52:07 -08:00
George Hotz
97fd9c1237 zero_grad there to match readme 2020-12-07 23:12:18 -08:00
George Hotz
c63f950348 need zero grad now 2020-12-07 23:10:43 -08:00
George Hotz
b355cd2571 Mean axis (doesn't work) (#154)
* mean axis

* fixed
2020-12-07 22:58:34 -08:00
George Hotz
38f97c8c80 prepare for ops_ane 2020-12-07 21:54:22 -08:00
George Hotz
7f249ec76d touch up 2020-12-07 21:51:32 -08:00
Marcel Bischoff
58ccebd7cd Sum with axis (#153)
* sum with axis and tests

* broken

* works again

* clean up

* Update test_ops.py
2020-12-07 21:49:18 -08:00
George Hotz
ac9fecb05d lots of notes 2020-12-07 21:40:31 -08:00
George Hotz
8d1500f497 conv neuron 2020-12-07 21:12:52 -08:00
George Hotz
e4bb53b0e9 work out more 2020-12-07 20:32:50 -08:00
George Hotz
4927ad1897 float16 weights in min.weights 2020-12-07 20:15:15 -08:00
George Hotz
3aac9aefce fix GPU profiling 2020-12-07 20:03:28 -08:00
James Roberts
b2eca6d45f Format debug output (#152) 2020-12-07 14:07:14 -08:00
George Hotz
c7973cb0a1 ugh buffer_np is bad 2020-12-07 08:07:00 -08:00
George Hotz
088f280dc3 touchups 2020-12-07 07:50:27 -08:00
George Hotz
0cf21881b7 hwx parse w/o macho mods 2020-12-06 23:13:28 -08:00
Josh Smith
aa4161f63e use classmethods for Tensor helper funcs (#146) 2020-12-06 22:35:43 -08:00
George Hotz
e75a6d1b4b quadconv 2020-12-06 20:39:50 -08:00
George Hotz
23664c99bd double conv 2020-12-06 20:26:02 -08:00
George Hotz
1a0f826dc6 highlight the commands 2020-12-06 20:03:21 -08:00
George Hotz
2f1f006003 we have docs 2020-12-06 19:54:03 -08:00
George Hotz
3531e81f0e dumping ANE docs 2020-12-06 18:58:36 -08:00
George Hotz
6e793e96c3 deeebug 2020-12-06 17:49:17 -08:00
George Hotz
dced0cb44b oops, path to weights 2020-12-06 16:33:42 -08:00
George Hotz
c57dc61ea7 simple op examples 2020-12-06 16:32:26 -08:00
George Hotz
7babf38617 found concat 2020-12-06 16:27:12 -08:00
George Hotz
ddd6778423 add neuron 2020-12-06 16:24:42 -08:00
George Hotz
f2f2d6aea3 docs and noop 2020-12-06 16:10:44 -08:00
George Hotz
6ba25834ee found some plists 2020-12-06 15:57:28 -08:00
George Hotz
e2184c20ad min weights, update golden 2020-12-06 15:29:15 -08:00
George Hotz
d4d8bd0337 make minimal plist for compare 2020-12-06 15:10:15 -08:00
George Hotz
0845ec43c6 compile takes in plist 2020-12-06 14:51:33 -08:00
George Hotz
00312b8ad1 batchnorm work 2020-12-06 14:40:07 -08:00
George Hotz
da514c2918 fix enet init 2020-12-06 13:52:07 -08:00
George Hotz
3b982f2f7a get_parameters 2020-12-06 13:47:28 -08:00
George Hotz
102e6356e9 replace layer_init_uniform with .uniform 2020-12-06 13:44:31 -08:00
George Hotz
c71a8ef222 remove unused Div op 2020-12-06 13:02:12 -08:00
George Hotz
20f95de408 less lines 2020-12-06 12:56:03 -08:00
George Hotz
629efb391f lose a few lines 2020-12-06 12:42:41 -08:00
George Hotz
521098cc2f se optional, track time better 2020-12-06 12:29:42 -08:00
George Hotz
609d11e699 trainer works with CIFAR 2020-12-06 12:20:14 -08:00