George Hotz
69ca7f7bf9
changes for teenygrad ( #3665 )
...
* changes for teenygrad
* upd
* simpler test
2024-03-09 15:30:34 -08:00
chenyu
8f10bfa2ff
ban __bool__ on Tensor ( #3632 )
...
* ban __bool__ on Tensor
avoid misuse
* test case
* fix tests
* fix more tests
2024-03-06 17:12:35 -05:00
chenyu
968d109453
apply more create_lt_node ( #3597 )
...
updated one in linearizer if condition, and various symbolic tests
2024-03-03 16:12:39 -05:00
Patrick Tsai
0082300a59
Fix symbolic negative floordiv ( #3594 )
...
Co-authored-by: Patrick Tsai <patosai@users.noreply.github.com >
2024-03-03 11:40:52 -05:00
chenyu
e09619ab6c
explicitly create_lt_node when used in shapetracker _expr_view ( #3561 )
...
* explicitly create_lt_node when used in shapetracker
leave regular __lt__ and cmps for symbolic shape cmp
* hmm it fixed that?
* LtNode.substitute uses create_lt_node
2024-03-03 10:08:21 -05:00
Marcin Słowik
f90caa4b92
Escape table name in diskcache queries. ( #3543 )
...
Some devices create cache table names with non-alphanumerical characters, e.g. "compile_hip_gfx1010:xnack-_12".
This commit escapes the table name in single quotes s.t. sqlite works (see https://github.com/tinygrad/tinygrad/issues/3538 ).
2024-02-29 13:04:21 -08:00
George Hotz
48918fa75a
fix disktensor offset issue ( #3532 )
2024-02-28 17:22:17 -08:00
chenyu
0c6846f9fc
failed test case for disk tensor assign into dtype int64 ( #3527 )
...
failed case for #3510 , mark as expectedFailure for now
2024-02-28 17:52:21 -05:00
chenyu
88939c3347
fix Node.max can be symbolic ( #3514 )
...
Also made sure taking max twice can get int.
2024-02-27 17:21:31 -05:00
chenyu
61605ccc69
Remove special case of SumNode div SumNode ( #3502 )
2024-02-26 09:42:06 -05:00
George Hotz
871ba73e65
_reduce_op is axis based now ( #3462 )
...
* _reduce_op is axis based now
* axis_
* update lin failures
* disable that
* fix shape
2024-02-21 16:36:31 +01:00
chenyu
0d326a48b8
fix LtNode simplification when lhs and rhs contain same variables ( #3451 )
...
* fix LtNode simplification when lhs and rhs contain same variables
`(Variable("a", 1, 5) < Variable("a", 1, 5))` should eval to `NumNode(0)`
* fix with less perf impact
2024-02-20 09:06:55 -05:00
chenyu
2da734920e
use __getnewargs__ to fix unpickling Variable ( #3441 )
...
it's recommended to use __getnewargs__ to update the args of classes that use __new__ when unpickling.
It's preferred because it does not change the __new__ behavior.
2024-02-18 10:28:37 -05:00
xarkes
28a8b72024
Remove Interpreted device & remaining CPU/TORCH ref ( #3423 )
...
* Remove Interpreted device & remaining CPU/TORCH ref
* Oops
* supports_device was useful
* Fix doc wording
---------
Co-authored-by: chenyu <chenyu@fastmail.com >
2024-02-16 00:30:21 -05:00
George Hotz
93eceef727
remove cpu prereqs ( #3410 )
2024-02-15 13:45:06 +01:00
Jyotirmaya Mahanta
b6a2600c86
fix merging condition in merge_dims ( #3363 )
...
* fix merging condition in merge_dims
* add tests
* set contiguous after mask is canonicalized
* minor fix
2024-02-12 11:50:26 +01:00
chenyu
97275101e9
fix safetensor load uint32 and uint64 ( #3315 )
...
the correct keys are U32 and U64.
2024-02-04 10:46:27 -05:00
Yoshinori Sano
edb74897b2
support safe load bf16 ( #3310 )
...
* support safe load bf16
* fix lint error E501
* add test for loading safetensors
* key should be BOOL
* fix lint
2024-02-04 10:08:39 -05:00
chenyu
d459956966
move TestGetContraction to test_helpers ( #3313 )
...
also cleaned long lines in test_shapetracker and enabled the line length check
2024-02-04 06:05:01 -05:00
Felix Wu
021eea3a52
fix UnboundLocalError when running Compiler with DISABLE_COMPILER_CACHE ( #3296 )
2024-02-01 21:12:33 -05:00
David Hou
3378625773
name upcast variables ( #3200 )
...
* name upcast variables
* typing
* unused
2024-01-22 11:37:28 -05:00
George Hotz
d2aab65958
remove unused expr node ( #3170 )
...
* remove unused expr node
* still works
* simple expr_idxs
* fixup typing
2024-01-18 14:18:43 -08:00
George Hotz
f0c178b7e9
move get_contraction to helpers ( #3162 )
...
* move get_contraction to helpers
* move simplify
* lines
* to_movement_ops is not generic
2024-01-17 19:13:11 -08:00
George Hotz
a464909d79
fast resnet eval ( #3135 )
...
* fast resnet eval
* fix HIP multidevice graph
* neater expression for devices
* lines
* add decorator test
2024-01-15 14:15:18 -08:00
Paul Gustafson
6bb65cd02e
fix off-by-one error in st_equal ( #3131 )
...
* fix off by one error
* whitespace
2024-01-15 11:32:13 -08:00
chenyu
c658aa4fbf
minor cleanup of test_disk_tensor ( #3112 )
2024-01-13 20:54:58 -05:00
chenyu
a300fea2a4
failed test case due to cast resets shapetracker ( #3109 )
...
cast implicitly resets shapetracker and makes it contiguous (for disk tensor), which fails for Interpreted backend if inputs contain non-contiguous st.
2024-01-13 12:46:51 -05:00
chenyu
f018a55ea1
update NumNode.__hash__ to be hash(self.b) ( #3105 )
...
with this, `a:=NumNode(x) == b` implies `hash(a) == hash(b)`
2024-01-12 19:46:21 -05:00
chenyu
dab8214103
unit tests for Device.canonicalize ( #3055 )
2024-01-09 12:47:20 -05:00
George Hotz
655c6f61d3
St real size ( #3046 )
...
* track the size in the lazybuffer
* shapetracker real size
* lint
2024-01-08 14:44:53 -08:00
George Hotz
c003be7309
Revert "track size in shapetracker" ( #3043 )
...
* Revert "track size in shapetracker (#3026 )"
This reverts commit a8ba1ac08f .
* st.size
2024-01-08 13:13:39 -08:00
George Hotz
a8ba1ac08f
track size in shapetracker ( #3026 )
...
* track size in shapetracker
* shapetracker adapter
* size is an int
* create Buffer with st.size
* only compare the views for the jit
* fix webgpu
2024-01-05 20:15:53 -08:00
George Hotz
60abc62a3f
fast hip read ( #3014 )
...
* fast hip read
* hip read faster
* fix tests
* to_mv
* simplify
* bump to 6k lines
2024-01-05 10:33:13 -08:00
George Hotz
9699c8c90b
don't alloc for InterpretedASTRunner ( #2999 )
2024-01-03 17:05:53 -08:00
chenyu
74cc6fd3c2
remove AndNode.__floordiv__ special case ( #2996 )
...
* remove AndNode.__floordiv__
AndNode produces a Node that min/max is bounded by [0, 1] so `//` on top of that is almost always 0.
we don't really use that either
* keep the test
2024-01-03 17:44:55 -05:00
chenyu
ff5399f053
move one last dtype test from test_helpers to test_dtype ( #2975 )
2024-01-02 12:37:56 -05:00
George Hotz
a280cfe169
move dtypes to dtype.py ( #2964 )
...
* move dtypes to dtype.py
* fix urllib
2024-01-01 14:58:48 -08:00
George Hotz
c81ce9643d
move globalcounters to ops ( #2960 )
...
* move globalcounters to ops
* missed a few
* sick of that failing
2024-01-01 14:21:02 -08:00
chenyu
8291986959
Variable.sum -> Node.sum, Variable.ands -> Node.ands ( #2961 )
2024-01-01 16:21:28 -05:00
chenyu
3d720b5761
move expand_idx, iter_idxs and expand_node from symbolic to linearizer ( #2959 )
2024-01-01 14:41:21 -05:00
George Hotz
5cac6338a4
apply the multitensor optimizations in lazy.py ( #2901 )
...
* apply the multitensor optimizations in lazy.py
* less lines
* hack for webgpu
* save a line
2023-12-21 13:55:49 -08:00
George Hotz
1765849937
new lazy, benchmark ( #2878 )
...
* lazy rewrite, try 2
* min fix tests
* pass contig test
* put broken pads back
* move that to realize
* no contig child fixes array packing
* so wrong
* now that's correct
* base children
* fix bind issues
* disable to_image_idx
* fix tests
* that failure shouldn't break other tests
* more fixes
* fix torch
* skip failing tests in CI
* 1e-7
* half is broken
* 1e-6 margin of error
2023-12-20 14:33:21 -08:00
Peter Cawley
dae8976889
Fix reshape merging with masks ( #2877 )
2023-12-20 14:00:58 -08:00
George Hotz
ca59054463
fix shapetracker math ( #2861 )
...
* proper test
* all st math good now
* fix real_strides bug
2023-12-19 22:17:34 -08:00
chenyu
5a739e8c20
update one skipped pad_reshape test that was fine ( #2860 )
...
* update one skipped pad_reshape test that was fine
had a typo
* this one passed
2023-12-19 23:25:52 -05:00
chenyu
ad233d557f
disable reshape merging with masks ( #2858 )
...
fuzzer found a bug, and it's not complete
2023-12-19 19:06:16 -05:00
Oleg Rybalko
42a038c83f
More readable torch_load ext check ( #2853 )
...
* more readable extension check
* enable tarfile test
* detach tensor if requires grad in torch
2023-12-19 14:53:15 -05:00
George Hotz
b2192b5400
minor improvements ( #2845 )
2023-12-18 22:09:08 -08:00
George Hotz
d086325b1b
hotfix: failing tests
2023-12-18 21:12:42 -08:00
George Hotz
b6d71b131e
hotfix: push broken tests
2023-12-18 21:08:42 -08:00