Release v0.4.1

This commit is contained in:
Alex O'Connell
2025-10-12 10:37:46 -04:00
parent 5cb910d71c
commit e28132a17a
4 changed files with 4 additions and 4 deletions

View File

@@ -158,6 +158,7 @@ python3 train.py \
## Version History
| Version | Description |
|---------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| v0.4.1 | Fix an issue with using Llama.cpp models downloaded from HuggingFace |
| v0.4 | Rewrite integration to support tool calling models/agentic tool use loop, voice streaming, multiple config sub-entries per backend, and dynamic llama.cpp processor selection |
| v0.3.11 | Bug-fixes and llama.cpp version update |
| v0.3.10 | Add support for the OpenAI "Responses" API endpoint, Update llama.cpp version, Fix for breaking change in HA version 2025.7.0 |

View File

@@ -1074,8 +1074,7 @@ class LocalLLMSubentryFlowHandler(ConfigSubentryFlow):
storage_folder = os.path.join(self.hass.config.media_dirs.get("local", self.hass.config.path("media")), "models")
async def download_task():
# return await self.hass.async_add_executor_job(
await self.hass.async_add_executor_job(
return await self.hass.async_add_executor_job(
download_model_from_hf, model_name, quantization_type, storage_folder
)

View File

@@ -317,5 +317,5 @@ OPTIONS_OVERRIDES = {
},
}
INTEGRATION_VERSION = "0.4.0"
INTEGRATION_VERSION = "0.4.1"
EMBEDDED_LLAMA_CPP_PYTHON_VERSION = "0.3.16+b6153"

View File

@@ -1,7 +1,7 @@
{
"domain": "llama_conversation",
"name": "Local LLMs",
"version": "0.4.0",
"version": "0.4.1",
"codeowners": ["@acon96"],
"config_flow": true,
"dependencies": ["conversation"],