George Hotz
411392dfb7
move files into uop dir ( #10399 )
...
* move files into uop dir [pr]
* tinygrad.uop is a thing
* fix uop docs, no pr
* fix viz
2025-05-18 11:38:28 -07:00
chenyu
f6cf25fce4
cleanup test_conv2d_ceildiv_edge_case [pr] ( #10317 )
2025-05-14 23:35:28 -04:00
Kirill R.
50d7162acd
Add conv2d ceildiv edge case ( #10303 )
2025-05-14 22:50:23 -04:00
chenyu
17d4d258ea
simple symbolic slice in llama [pr] ( #10112 )
...
support slice that has step None and stop > start
2025-04-30 14:36:35 -04:00
George Hotz
a2d0684fc1
test_attention_simple_view ( #10092 )
...
* test_attention_simple_view
* correct comment
2025-04-28 20:01:22 -04:00
George Hotz
d32f5e9f3a
improve rendering of shapes in viz + investigate symbolic [pr] ( #10091 )
2025-04-28 16:44:09 -04:00
Sieds Lykles
e75be6eafc
[bounty] [pr] index validation with z3 ( #9981 )
...
* index validation with z3
* Change comment
* toposort -> toposort()
---------
Co-authored-by: George Hotz <72895+geohot@users.noreply.github.com >
2025-04-24 08:06:08 -04:00
chenyu
f05fd118a2
few minor code cleanups [pr] ( #8267 )
2024-12-15 23:44:51 -05:00
George Hotz
5ae2de9845
UOp.variable ( #7010 )
...
* UOp.variable [pr]
* fix tests
* clean
* improve name rendering
* last bug
2024-10-12 18:20:44 +08:00
chenyu
2de174677a
threefry touchup [run_process_replay] ( #6169 )
...
also why is test_gc testing _rng_counter is allocated??
2024-08-18 23:01:24 -04:00
chenyu
a352b6d9ce
symbolic Tensor.var ( #4843 )
...
taken from #4446 and add more tests
2024-06-05 12:55:54 -04:00
chenyu
236390aafb
fix lazy r const folding with variable shape ( #4783 )
...
currently not supporting const fold symbolic shape. I think it's possible with a refactor to Tensor.from_node.
also added some failed required tests for symbolic arange.
2024-05-30 15:19:28 -04:00
chenyu
7fab8c9e17
add symbolic mean test cases in test_symbolic_ops and test_symbolic_jit ( #4523 )
...
* add symbolic mean test cases in test_symbolic_ops and test_symbolic_jit
2d symbolic mean in jit does not quite work, order of the variable inputs are not deterministic?
* skip
2024-05-10 23:19:55 -04:00
George Hotz
81baf3eed3
bring ptx back ( #3623 )
...
* bring ptx back
* ptx back
* fix define var
* fix a few bugs
* bugfixes
* fixes
* fix llvm bug
* fix test bug
2024-03-06 13:34:21 -08:00
chenyu
969b57f0fe
enable symbolic_ops and jits test of two vars ( #3513 )
2024-02-27 11:17:46 -05:00
chenyu
f96fc6e9d4
fix gpt2 with empty prompt take 2 ( #3102 )
...
logits would be empty so need to replace that with ones before sampling, also cannot reshape with -1 when there's 0 in other axes
2024-01-12 14:46:36 -05:00
chenyu
ca46d3541b
Revert "fix gpt2 with empty prompt" ( #3101 )
2024-01-12 14:27:41 -05:00
chenyu
1d7f01bc6d
fix gpt2 with empty prompt ( #3100 )
...
logits would be empty so need to replace that with ones before sampling, also cannot reshape with -1 when there's 0 in other axes
2024-01-12 14:18:17 -05:00
chenyu
f0d7ad8aaa
fix gpt2 attention with start_pos = 0 ( #3061 )
...
* fix gpt2 attention with start_pos size 1
test cases taken from ll_transformer branch
* fix interpreted
2024-01-09 16:14:55 -05:00
George Hotz
1765849937
new lazy, benchmark ( #2878 )
...
* lazy rewrite, try 2
* min fix tests
* pass contig test
* put broken pads back
* move that to realize
* no contig child fixes array packing
* so wrong
* now that's correct
* base children
* fix bind issues
* disable to_image_idx
* fix tests
* that failure shouldn't break other tests
* more fixes
* fix torch
* skip failing tests in CI
* 1e-7
* half is broken
* 1e-6 margin of error
2023-12-20 14:33:21 -08:00
chenyu
e13b4964d7
remove the all_int(shape) check in Tensor._loadop ( #2874 )
...
* remove the all_int(shape) check in Tensor._loadop
we can support jittable symbolic shape random with custom rand now, and we can formalize it in the test after threefry is ready
* MOCKHIP false positive
2023-12-20 15:04:50 -05:00
Ahmed Harmouche
4b01839774
support vals on WebGPU, run more tests ( #2668 )
...
* Vals on webgpu, run more tests
* Skip slow tests, run symbolic ops tests
* Balance out tests
2023-12-07 16:45:21 -08:00
chenyu
28a67106ca
enable symbolic ops tests for hip ( #2485 )
2023-11-27 22:33:41 -08:00
chenyu
9eeba968cd
fix the variable arg order ( #2382 )
2023-11-21 12:02:31 -05:00
George Hotz
70a65c201e
JIT support in Interpreted ( #2314 )
...
* factor that out
* jit is supported everywhere
* fix some tests
* there's no jit supported device, the jit is everywhere
* fix test uops
2023-11-15 11:13:38 -08:00
chenyu
d86ea188dd
support symbolic shape in Interpreted ( #2289 )
...
* support symbolic shape in Interpreted
* simpler
* no InterpretedFlopCounter
* tragic NumNode
* regex is hard
2023-11-13 20:13:18 -05:00
chenyu
e2b83f1b42
Variable.bind newer ( #2017 )
...
* Variable.bind attempt 2
* ShapeTracker.unbind
* fix llama
* fix types
* test case
* View.vars cleanup
* include mask in symbolic source
* mask can be sint
* st.unbind in bufferops
* assert ast contain free Variable only
* cleanup
* conservative unbinding reduce op arg
* move reduceop unbind
* fix llama JIT arg behavior
2023-10-10 10:03:01 -07:00
Yixiang Gao
094d3d71be
with Tensor.train() ( #1935 )
...
* add with.train
* remove the rest TODOs
* fix pyflake
* fix pyflake error
* fix mypy
2023-09-28 18:02:31 -07:00
Gijs Koning
767bb35903
Enable symbolic ops tests for LLVM ( #1898 )
...
* Enable symbolic tests for HIP and LLVM
* Only llvm
2023-09-23 07:30:26 +08:00
chenyu
e67306ba04
symbolic shape type with TypeGuard ( #1852 )
2023-09-13 05:27:22 +08:00
chenyu
ebcda8a714
Move var_vals from ShapeTracker to LazyBuffer ( #1819 )
2023-09-08 09:25:10 -07:00
chenyu
66fbf4800b
fix symbolic_ops tests with Tensor.training=True ( #1686 )
2023-08-26 23:19:56 -04:00
chenyu
89e13f2f04
support symbols in shrink ( #1611 )
2023-08-22 09:08:21 -07:00
Yixiang Gao
8d6662a741
.cpu().numpy() -> .numpy() ( #1594 )
...
* .cpu().numpy() -> .numpy()
* restore ops_torch
* restore test_speed_v_torch
2023-08-21 09:53:29 -07:00
George Hotz
739f327d2d
Shorter ( #1582 )
...
* deleting lines
* remove insert dims
* if statement is never hit
* bug fixes
2023-08-20 08:12:16 -07:00
chenyu
dfec16cc83
Support arg int for CUDA kernel ( #1565 )
2023-08-18 09:19:40 -07:00
chenyu
11dd9b1741
symbolic codegen and exec ( #1552 )
...
* symbolic codegen and exec
* fix and add test
* no sketchy
* merge_dicts type
* dtypes._arg_int32
2023-08-16 14:43:41 -07:00