George Hotz
168c16646a
change create_schedule_with_vars api to big_sink [pr] ( #8677 )
2025-01-19 13:30:26 -08:00
ignaciosica
d2234e308a
tf32 tc for nv and ptx ( #8635 )
...
Co-authored-by: George Hotz <72895+geohot@users.noreply.github.com >
2025-01-17 17:43:57 -08:00
nimlgen
b3efeeb717
docs: start am docs ( #8638 )
...
* docs: init am docs
* missing
2025-01-16 00:22:35 +03:00
qazal
0e97f807e0
test fixup prereqs for delete_buffer_view [pr] ( #8523 )
2025-01-07 11:52:18 +02:00
nimlgen
5cb9443ebb
PROFILE is enabled when VIZ is enabled ( #8516 )
2025-01-06 19:47:16 +03:00
uuuvn
5ffc50d58c
Clang JIT ( #8481 )
...
Co-authored-by: George Hotz <72895+geohot@users.noreply.github.com >
2025-01-03 11:12:55 -05:00
qazal
bd4d7dc4eb
return becomes_map from the scheduler ( #8483 )
...
* return becomes_map from the scheduler
* fix test_schedule
* fix abstractions2
* s/becomes/becomes_map
2025-01-03 22:47:21 +08:00
chenyu
f3fdec940d
Tensor.mod ( #8458 )
...
it's a python style mod. possibily can be cleaner with a floor div
relaxed the vmin for MOD slightly for cstyle negatives mod, it's more correct and might fix other bugs
2024-12-31 11:31:42 -05:00
George Hotz
4c94726bac
remove uop mutability [pr] ( #8441 )
...
* remove uop mutability [pr]
* test fixups
* most tests pass
* more tests pass
* lil test fixups
* them too
* fix test
* unneeded
* err, that
* fix test_hcq
* fix test failures
* fix that test
* tensor universe
* does this pass test
* Revert "does this pass test"
This reverts commit ed516b3169 .
* Revert "tensor universe"
This reverts commit c21301852a .
* proper spidering for uops
* cleanups
* all tensors
* all tensors
* slow but correct
* fast
* no WeakSet
* faster
* no need for list
* revert that
2024-12-31 00:29:56 -05:00
chenyu
19a54ae0b4
add Tensor.roll and Tensor.rearrange to doc ( #8454 )
...
also moved rearrange in tensor.py to high level movement
2024-12-30 20:25:50 -05:00
George Hotz
803a47494e
Revert "Clang JIT ( #8312 )" ( #8452 )
...
This reverts commit b6266c8e41 .
2024-12-30 17:49:35 -05:00
uuuvn
b6266c8e41
Clang JIT ( #8312 )
...
Co-authored-by: George Hotz <72895+geohot@users.noreply.github.com >
2024-12-30 17:37:53 -05:00
qazal
866dfa1f23
create_schedule([x.lazydata]) -> x.schedule() in tests ( #8449 )
2024-12-31 03:15:52 +08:00
geohotstan
78cb47dfc5
docs and tests clean ups ( #8383 )
2024-12-23 11:12:13 -05:00
chenyu
63f195729d
add gguf_load to doc [pr] ( #8314 )
...
Co-authored-by: George Hotz <72895+geohot@users.noreply.github.com >
2024-12-18 12:44:09 -05:00
qazal
d05e21cb69
replace lazy srcs with the new uop api [pr] ( #8255 )
...
* buf_uop_view function
* srcs shouldn't exist
* fix TestTensorMetadata
---------
Co-authored-by: George Hotz <geohot@gmail.com >
2024-12-15 17:09:54 +08:00
George Hotz
8396d90f91
non controversial changes from optim branch [pr] ( #8234 )
2024-12-13 19:24:16 -08:00
George Hotz
37fa38d272
Revert "switch beautiful_mnist to use new optimizer [pr] ( #8231 )" ( #8233 )
...
This reverts commit e9ee39df22 .
2024-12-13 19:07:09 -08:00
George Hotz
e9ee39df22
switch beautiful_mnist to use new optimizer [pr] ( #8231 )
...
* switch beautiful_mnist to use new optimizer [pr]
* fix abstractions3 + docs
* fix OptimizerGroup with schedule_step api
2024-12-13 18:27:16 -08:00
George Hotz
8a04a3a77a
rename LazyBuffer -> UOp [pr] ( #8169 )
...
* rename LazyBuffer -> UOp [pr]
* fix docs
2024-12-11 16:15:52 -08:00
geohotstan
0a2e10be1d
add SELU to Tensor ( #7993 )
...
* add selu
* more clean ups
2024-12-02 10:04:01 -05:00
nimlgen
10f431b96d
hcq replace update with sint ( #7899 )
...
* try sym hcq
* start with amd
* move to nv
* nv works
* cache and qcom
* fixes
* signals
* fix nv
* qcom fixes
* linter
* linter
* cache + typings
* fixes
* tiny fixes
* linter
* linter
* lntr
* ugh
* comments
2024-11-29 20:08:13 +03:00
geohotstan
cea5853cfa
add Tensor.scatter ( #7737 )
...
* working I think
* where are my onnx scatter tests??
* forward_only for now
* try if nan hack fix NV
* looks like issue is different... CUDA WHY
* oops that was wrong. Try if this fixes CUDA
* simpler multiply
* actually finish this up tmrw morning :x
* fix tests?
* improve tests
* improve test and implementation
* fix ruff
* complete but lots of expected failure...
* reviewed tests
* add onnx tests
* is this a processing op?
* add return type to indicate that it's not in-place
* final cleanups
* use or and improve tests a little
* add masked_index_select
* call it masked_setitem instead
* try
* FIXED
---------
Co-authored-by: chenyu <chenyu@fastmail.com >
2024-11-27 10:52:04 -05:00
chenyu
3b26e51fce
Tensor.cummax ( #7854 )
...
generalized the existing cumsum and take Ops.MAX in addition to Ops.ADD
2024-11-22 15:55:02 -05:00
geohotstan
cf1ec90ad4
add inverse trig functions to Tensor ( #7805 )
...
* implement inverse trig functions
* guess we should still test nans?
* magnitude as variable name :D
* reorder onnx_ops ops
* approximation -> x for consistency
* address feedback
* simpler acos
* improvement?
* actually just have asin depend on atan
* actually this is nicer
* remove a comment
---------
Co-authored-by: chenyu <chenyu@fastmail.com >
2024-11-21 09:13:36 -05:00
George Hotz
9df5a62c5e
unify to HWQueue [pr] ( #7812 )
...
* unify to HWCommandQueue [pr]
* all is HWQueue
2024-11-21 10:33:08 +08:00
George Hotz
d71fe7faa5
rename allocator methods to not conflict [pr] ( #7788 )
...
* rename allocator methods to not conflict [pr]
* forgot those
* transfer + offset
2024-11-20 00:10:29 +08:00
geohotstan
72a41095bc
add Tensor.meshgrid ( #7714 )
...
* initial implementation and test
* some other places that can use meshgrid
* revert the onnx_ops change
* add to docs
* revert interpolate too
* update
* improve edge case test
* might as well test grad
* add to test can improve docs
---------
Co-authored-by: chenyu <chenyu@fastmail.com >
2024-11-16 23:06:47 -05:00
ignaciosica
597a239e28
Remove UnaryOps, BinaryOps, TernaryOps, MetaOps [pr] ( #7725 )
...
* remove unaryops
* remove ternaryops
* remove metaops
* hotfix
* remove binaryops
* hotfix: test_pattern_matcher
---------
Co-authored-by: qazal <77887910+Qazalin@users.noreply.github.com >
2024-11-16 20:56:56 +08:00
geohotstan
f8056a74d6
combine pad2d with pad ( #7677 )
...
* I have pad2d, I have pad, uuh~, pad2dpad~
* fix some small things
* strategically placed cast hack
* fix more
* fix more more
* tests
* periods
2024-11-14 17:56:02 +08:00
chenyu
51afc3cc88
update env_vars doc on VIZ link ( #7689 )
...
existing one throws 404 because mkdocs does not allow traverse above doc root (i think?). so for now just stick the github link to it
2024-11-13 17:28:14 -05:00
geohotstan
9c41c376d3
add Tensor.nll_loss ( #7683 )
...
* move nll_loss to new branch
* make nll_loss examples practical
* self *is*
* add to docs
* small
2024-11-13 13:12:13 -05:00
geohotstan
5eef59d732
add Tensor.linspace ( #7609 )
...
* add linspace
* shave off tests and forgot to add to docs crap
* WHOOPS
* better tests
2024-11-12 10:29:36 +08:00
Kinvert
6a0ed46b1c
adding viz to env_vars docs ( #7630 )
2024-11-11 21:28:27 +08:00
George Hotz
c8bf09b7d4
s/UOps/Ops ( #7500 )
...
* s/UOps/Ops [pr]
* fix
2024-11-03 11:26:10 +08:00
geohotstan
585f3a0f24
Add isinf and isnan ops to Tensor ( #7484 )
...
* move isinf and isnan to new branch
* sneak a roll documentation fix in
* add to docs
* update test coverage for detect_positive and detect_negative
* add types to isinf args
2024-11-02 12:12:52 -04:00
geohotstan
6513690223
Add Tensor.hardsigmoid ( #7433 )
...
* move hardsigmoid to new branch
* add to test
* add NOTE to mention differing values for alpha and beta that match torch
* shift from relu6
* correct shift implementation
* or we just use relu? no more 666
2024-11-01 08:36:52 -04:00
chenyu
fb694a63eb
Tensor.erf ( #7419 )
...
the same one used in onnx and the one in bert.
2024-10-30 18:12:28 -04:00
vinzentbeer
573a848229
fix small typo ( #7399 )
...
"We use with Tensor.train() set the internal flag" -> "We use with Tensor.train() *to* set the internal flag"
2024-10-30 19:20:28 +08:00
chenyu
96fcc47e27
touchup abstraction docs ( #7327 )
...
fix typing and use tinygrad tqdm
2024-10-27 22:29:55 -04:00
George Hotz
4438d6a467
Tensor.from_url API [pr] ( #7210 )
...
* Tensor.fetch API [pr]
* update docs
* from_url
2024-10-22 14:54:17 +08:00
George Hotz
ded1b38b84
minor dtype cleanup [pr] ( #7124 )
...
* minor dtype cleanup [pr]
* use ptr() function
2024-10-17 17:41:23 +08:00
George Hotz
3169cb386d
remove graph [pr] ( #7085 )
2024-10-16 11:40:07 +08:00
Harsh Natuskar
ace834ef7b
=docs update ( #7027 )
2024-10-13 19:39:06 +08:00
jeffzh4ng
19a7e41113
implement logcumsumexp ( #6921 )
...
* implement logcumsumexp
* change axis=None to axis=0
2024-10-06 10:45:36 -04:00
George Hotz
f588169fdc
hotfix: ad for DEBUG=2 in the mnist tutorial
2024-10-06 21:05:48 +08:00
George Hotz
4df5c7a4ef
move lazy to engine [pr] ( #6886 )
...
* move lazy to engine [pr]
* engine.lazy
2024-10-04 23:19:26 +08:00
George Hotz
6b063450df
move hcq device to runtime [pr] ( #6879 )
...
* things that are only used in one place don't belong in helpers [pr]
* start moving hcq device [pr]
* fix paths
2024-10-04 22:26:50 +08:00
nimlgen
3c56aeee70
add Tensor.from_blob ( #6765 )
...
* draft tensor from pointer init
* some docs and types
* comment
* cleaner
* test
* malloc
* qcom cl interop
* jit example
* cleaner
* dealoc
* wording
* docs
2024-09-26 18:33:19 +08:00
George Hotz
e015b41ce9
remove e( function just alu( [run_process_replay] ( #6589 )
...
* remove e( function just alu( [run_process_replay]
* missed two
2024-09-19 10:24:02 +08:00