From 787671c2c29b519b8e28feb1bb82df3170a688a2 Mon Sep 17 00:00:00 2001 From: Lincoln Stein Date: Fri, 21 Jun 2024 15:15:31 -0400 Subject: [PATCH] Update invokeai/app/api/routers/model_manager.py Co-authored-by: Ryan Dick --- invokeai/app/api/routers/model_manager.py | 7 ++++++- 1 file changed, 6 insertions(+), 1 deletion(-) diff --git a/invokeai/app/api/routers/model_manager.py b/invokeai/app/api/routers/model_manager.py index 1730d383f0..ffb7e909e5 100644 --- a/invokeai/app/api/routers/model_manager.py +++ b/invokeai/app/api/routers/model_manager.py @@ -825,7 +825,12 @@ async def get_starter_models() -> list[StarterModel]: async def get_cache_size(cache_type: CacheType = Query(description="The cache type", default=CacheType.RAM)) -> float: """Return the current RAM or VRAM cache size setting (in GB).""" cache = ApiDependencies.invoker.services.model_manager.load.ram_cache - return cache.max_cache_size if cache_type == CacheType.RAM else cache.max_vram_cache_size + if cache_type == CacheType.RAM: + return cache.max_cache_size + elif cache_type == CacheType.VRAM: + return cache.max_vram_cache_size + else: + raise ValueError(f"Unexpected {cache_type=}.") @model_manager_router.put(