C T
0f9d7f650d
whisper: fix oob, explicit dtype ( #13144 )
...
* fix dtype depending on numpy version
numpy v2 np.array returns int64 which Tensor passed through for the
first decode call, swallowing the <|notimestamps|> token and corrupting
the sequence
* fix whisper OOB
global limit on whisper's context length
* enforce whisper max_tokens_to_sample (match openai)
local limit on max tokens decoded
2025-11-07 12:55:01 -05:00
chenyu
84d2d047ea
Tensor.pad_to and Tensor.shrink_to ( #12210 )
...
most of the time i want this instead of spelling out the args
also add more input validation to shrink
2025-09-16 12:24:55 -04:00
chenyu
b072be0e2d
hotfix whisper main script ( #11184 )
2025-07-11 12:34:00 -04:00
Nino Risteski
bc15e98f5c
clean up unused imports in examples and update CI linting ( #11024 )
...
* clean up unused imports in examples
* enable unused import checking in examples
* lint
* ignore F541 and F841 - focus on unused imports only
* clean up
* restore tinygrad.frontend.torch for TINY_BACKEND
* tiny change
2025-06-30 08:21:27 -07:00
Sieds Lykles
c29a56dd51
Fix whisper OOB ( #10685 )
...
* fix whisper and test
* remove import
2025-06-07 20:23:50 -04:00
kormann
f5dd25d376
enable whisper batch for long sequences ( #6458 )
...
* long batch +test
* long batch +test
* cleanup
* rollback syntactic changes
---------
Co-authored-by: chenyu <chenyu@fastmail.com >
2024-09-17 00:42:10 -04:00
kormann
f6f4f3222f
whisper long batch ( #6335 )
...
* reset
* test
* only part refactor
2024-09-09 21:03:59 -04:00
chenyu
a1940ced77
remove the assign hack in whisper ( #4240 )
...
no longer needed, the commented test case was removed too
2024-04-20 23:56:44 -04:00
George Hotz
150ea2eb76
create engine folder and move code ( #3948 )
...
* retry
* older tf
* that
2024-03-26 20:38:03 -07:00
George Hotz
41efaa848c
move graph.py and jit.py into features ( #3376 )
...
* move graph.py into features
* move jit into features
* fix quickstart
2024-02-12 17:34:34 +01:00
Francis Lata
6d672785db
Update Whisper to use fetch helper ( #2401 )
...
* update whisper to use new fetch helper
* simplify file opening
* update name
* update key name to "downloads-cache"
2023-11-23 12:59:59 -08:00
mmmkkaaayy
7f0cc4a4e8
whisper: support audio >30s ( #2378 )
...
* whisper: support audio >30s
* make prompt indexing consistent with reference repo
* fix online
2023-11-21 14:37:51 -08:00
chenyu
e9847be790
remove whisper +1-1 hack ( #2360 )
...
* remove whisper +1-1 hack
* Revert "remove whisper +1-1 hack"
This reverts commit 5db3800f09 .
* update whisper tests
* comment context
2023-11-19 17:56:36 -05:00
mmmkkaaayy
8235da11dd
whisper: support batch inference, add librispeech WER test ( #2074 )
...
* whisper: support batch inference, add librispeech WER test, add kv caching and JIT
* remove JIT_SUPPORTED_DEVICE
---------
Co-authored-by: George Hotz <72895+geohot@users.noreply.github.com >
2023-11-16 13:50:08 -08:00
mmmkkaaayy
91168a28c4
whisper: make file transcription work, add basic CI test ( #2042 )
2023-10-13 17:13:35 -07:00
mmmkkaaayy
af6e2f31ca
whisper: cast model output token to int32 ( #2013 )
...
Co-authored-by: mmmkkaaayy <mmmkkaaayy@users.noreply.github.com >
2023-10-08 05:56:22 -07:00
Karan Handa
a8aa13dc91
[ready] Replacing os with pathlib ( #1708 )
...
* replace os.path with pathlib
* safe convert dirnames to pathlib
* replace all os.path.join
* fix cuda error
* change main chunk
* Reviewer fixes
* fix vgg
* Fixed everything
* Final fixes
* ensure consistency
* Change all parent.parent... to parents
2023-08-30 10:41:08 -07:00
George Hotz
718ced296c
move state to nn/state ( #1619 )
2023-08-22 07:36:24 -07:00
Umut Zengin
f720682beb
np.argmax to Tensor.argmax ( #1608 )
...
* to tensor argmax
* removed keepdim
* training update
2023-08-21 15:22:29 -07:00
Paolo Gavazzi
9ffa1eb7e2
Removed dep of torch, torchaudio, kept librosa only ( #1264 )
2023-08-02 13:52:04 -04:00
cloud11665
e8a23d4331
there is a better way to do that! ( #950 )
2023-06-06 15:23:30 -07:00
George Hotz
b78addf2f8
Whisper ( #919 )
...
* no whispering yet
* whispering
* live whisper
* small support
2023-06-03 18:55:14 -07:00