Ryan Dick
497bc916cc
Add unet_config to get_scheduler(...) call in TiledMultiDiffusionDenoiseLatents.
2025-01-15 08:44:08 -05:00
dunkeroni
ebe1873712
fix: only add prediction type if it exists
2025-01-15 08:44:08 -05:00
dunkeroni
59926c320c
support v-prediction in denoise_latents.py
2025-01-15 08:44:08 -05:00
Mary Hipp
2d3e2f1907
use window instead of document
2025-01-14 20:01:08 -05:00
psychedelicious
d88b59c5c4
Revert "feat(ui): rearrange canvas paste back nodes to save an image step"
...
This reverts commit 7cdda00a54 .
2025-01-10 15:59:29 +11:00
psychedelicious
8da9d3bc19
chore: bump version to v5.6.0rc2
2025-01-09 14:12:46 +11:00
psychedelicious
a329588e5a
feat: add link to low vram guide to OOM toast (local only)
...
Needed to do a bit of refactoring to support this. Overall, the error toast components are easier to understand now.
2025-01-09 11:20:05 +11:00
psychedelicious
e09cf64779
feat: more updates to first run view
2025-01-09 11:20:05 +11:00
psychedelicious
e6deaa2d2f
feat(ui): minor layout tweaks for first run screen
2025-01-09 11:20:05 +11:00
psychedelicious
5246b31347
feat(ui): add low vram link to first run page
2025-01-09 11:20:05 +11:00
psychedelicious
89b576f10d
fix(ui): prevent canvas & main panel content from scrolling
...
Hopefully fixes issues where, when run via the launcher, the main panel kinda just scrolls out of bounds.
2025-01-09 09:14:22 +11:00
psychedelicious
d7893a52c3
tweak(ui): whats new copy
2025-01-08 15:26:26 +11:00
Mary Hipp
b9c45c3232
Whats new update
2025-01-08 15:26:26 +11:00
David Burnett
afc9d3b98f
more ruff formating
2025-01-07 20:18:19 -05:00
David Burnett
7ddc757bdb
ruff format changes
2025-01-07 20:18:19 -05:00
David Burnett
d8da9b45cc
Fix for DEIS / DPM clash
2025-01-07 20:18:19 -05:00
Ryan Dick
607d19f4dd
We should not trust the value of since the model could be partially-loaded.
2025-01-07 19:22:31 -05:00
Ryan Dick
974b4671b1
Deprecate the ram and vram configs to make the migration to dynamic
...
memory limits smoother for users who had previously overriden these
values.
2025-01-07 16:45:29 +00:00
Ryan Dick
85eb4f0312
Fix an edge case with model offloading from VRAM to RAM. If a GGML-quantized model is offloaded from VRAM inside of a torch.inference_mode() context manager, this will cause the following error: 'RuntimeError: Cannot set version_counter for inference tensor'.
2025-01-07 15:59:50 +00:00
psychedelicious
67e948b50d
chore: bump version to v5.6.0rc1
2025-01-07 19:41:56 +11:00
Riccardo Giovanetti
d9a20f319f
translationBot(ui): update translation (Italian)
...
Currently translated at 99.3% (1639 of 1649 strings)
Co-authored-by: Riccardo Giovanetti <riccardo.giovanetti@gmail.com >
Translate-URL: https://hosted.weblate.org/projects/invokeai/web-ui/it/
Translation: InvokeAI/Web UI
2025-01-07 19:32:50 +11:00
Riku
38d4863e09
translationBot(ui): update translation (German)
...
Currently translated at 71.7% (1181 of 1645 strings)
Co-authored-by: Riku <riku.block@gmail.com >
Translate-URL: https://hosted.weblate.org/projects/invokeai/web-ui/de/
Translation: InvokeAI/Web UI
2025-01-07 19:32:50 +11:00
Nik Nikovsky
cd7ba14adc
translationBot(ui): update translation (Polish)
...
Currently translated at 16.5% (273 of 1645 strings)
translationBot(ui): update translation (Polish)
Currently translated at 15.4% (254 of 1645 strings)
translationBot(ui): update translation (Polish)
Currently translated at 10.8% (178 of 1645 strings)
Co-authored-by: Nik Nikovsky <zejdzztegomaila@gmail.com >
Translate-URL: https://hosted.weblate.org/projects/invokeai/web-ui/pl/
Translation: InvokeAI/Web UI
2025-01-07 19:32:50 +11:00
Linos
e5b6beb24d
translationBot(ui): update translation (Vietnamese)
...
Currently translated at 100.0% (1649 of 1649 strings)
translationBot(ui): update translation (Vietnamese)
Currently translated at 100.0% (1645 of 1645 strings)
translationBot(ui): update translation (Vietnamese)
Currently translated at 100.0% (1645 of 1645 strings)
translationBot(ui): update translation (Vietnamese)
Currently translated at 100.0% (1645 of 1645 strings)
Co-authored-by: Linos <linos.coding@gmail.com >
Translate-URL: https://hosted.weblate.org/projects/invokeai/web-ui/vi/
Translation: InvokeAI/Web UI
2025-01-07 19:32:50 +11:00
Ryan Dick
d7ab464176
Offload the current model when locking if it is already partially loaded and we have insufficient VRAM.
2025-01-07 02:53:44 +00:00
Ryan Dick
548b3eddb8
pnpm typegen
2025-01-07 01:20:15 +00:00
Ryan Dick
5b42b7bd45
Add a utility to help with determining the working memory required for expensive operations.
2025-01-07 01:20:15 +00:00
Ryan Dick
71b97ce7be
Reduce the likelihood of encountering https://github.com/invoke-ai/InvokeAI/issues/7513 by elminating places where the door was left open for this to happen.
2025-01-07 01:20:15 +00:00
Ryan Dick
b343f81644
Use torch.cuda.memory_allocated() rather than torch.cuda.memory_reserved() to be more conservative in setting dynamic VRAM cache limits.
2025-01-07 01:20:15 +00:00
Ryan Dick
4abfb35321
Tune SD3 VAE decode working memory estimate.
2025-01-07 01:20:15 +00:00
Ryan Dick
cba6528ea7
Add a 20% buffer to all VAE decode working memory estimates.
2025-01-07 01:20:15 +00:00
Ryan Dick
6a5cee61be
Tune the working memory estimate for FLUX VAE decoding.
2025-01-07 01:20:15 +00:00
Ryan Dick
bd8017ecd5
Update working memory estimate for VAE decoding when tiling is being applied.
2025-01-07 01:20:15 +00:00
Ryan Dick
299eb94a05
Estimate the working memory required for VAE decoding, since this operations tends to be memory intensive.
2025-01-07 01:20:15 +00:00
Ryan Dick
fc4a22fe78
Allow expensive operations to request more working memory.
2025-01-07 01:20:13 +00:00
Ryan Dick
a167632f09
Calculate model cache size limits dynamically based on the available RAM / VRAM.
2025-01-07 01:14:20 +00:00
Ryan Dick
1321fac8f2
Remove get_cache_size() and set_cache_size() endpoints. These were unused by the frontend and refer to cache fields that are no longer accessible.
2025-01-07 01:06:20 +00:00
Ryan Dick
6a9de1fcf3
Change definition of VRAM in use for the ModelCache from sum of model weights to the total torch.cuda.memory_allocated().
2025-01-07 00:31:53 +00:00
Ryan Dick
e5180c4e6b
Add get_effective_device(...) utility to aid in determining the effective device of models that are partially loaded.
2025-01-07 00:31:00 +00:00
Ryan Dick
2619ef53ca
Handle device casting in ia2_layer.py.
2025-01-07 00:31:00 +00:00
Ryan Dick
bcd29c5d74
Remove all cases where we check the 'model.device'. This is no longer trustworthy now that partial loading is permitted.
2025-01-07 00:31:00 +00:00
Ryan Dick
1b7bb70bde
Improve handling of cases when application code modifies the size of a model after registering it with the model cache.
2025-01-07 00:31:00 +00:00
Ryan Dick
7127040c3a
Remove unused function set_nested_attr(...).
2025-01-07 00:31:00 +00:00
Ryan Dick
ceb2498a67
Add log prefix to model cache logs.
2025-01-07 00:31:00 +00:00
Ryan Dick
d0bfa019be
Add 'enable_partial_loading' config flag.
2025-01-07 00:31:00 +00:00
Ryan Dick
535e45cedf
First pass at adding partial loading support to the ModelCache.
2025-01-07 00:30:58 +00:00
Ryan Dick
c579a218ef
Allow models to be locked in VRAM, even if they have been dropped from the RAM cache (related: https://github.com/invoke-ai/InvokeAI/issues/7513 ).
2025-01-06 23:02:52 +00:00
Riku
f4f7415a3b
fix(app): remove obsolete DEFAULT_PRECISION variable
2025-01-06 11:14:58 +11:00
Mary Hipp
7d6c443d6f
fix(api): limit board_name length to 300 characters
2025-01-06 10:49:49 +11:00
psychedelicious
4815b4ea80
feat(ui): tweak verbiage for model install errors
2025-01-03 11:21:23 -05:00