From 9c8dd2576a156adc90f41a79f15427d1440bfa18 Mon Sep 17 00:00:00 2001 From: Alex O'Connell Date: Mon, 29 Jul 2024 21:42:13 -0400 Subject: [PATCH] Release v0.3.4 --- README.md | 1 + custom_components/llama_conversation/const.py | 2 +- custom_components/llama_conversation/manifest.json | 2 +- 3 files changed, 3 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 40b5865..0f20122 100644 --- a/README.md +++ b/README.md @@ -136,6 +136,7 @@ In order to facilitate running the project entirely on the system where Home Ass ## Version History | Version | Description | |---------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| +| v0.3.4 | Update bundled llama-cpp-python to support new models, various bug fixes | | v0.3.3 | Improvements to the Generic OpenAI Backend, improved area handling, fix issue using RGB colors, remove EOS token from responses, replace requests dependency with aiohttp included with Home Assistant | | v0.3.2 | Fix for exposed script entities causing errors, fix missing GBNF error, trim whitespace from model output | | v0.3.1 | Adds basic area support in prompting, Fix for broken requirements, fix for issue with formatted tools, fix custom API not registering on startup properly | diff --git a/custom_components/llama_conversation/const.py b/custom_components/llama_conversation/const.py index d9d011c..e8a625d 100644 --- a/custom_components/llama_conversation/const.py +++ b/custom_components/llama_conversation/const.py @@ -340,5 +340,5 @@ OPTIONS_OVERRIDES = { }, } -INTEGRATION_VERSION = "0.3.3" +INTEGRATION_VERSION = "0.3.4" EMBEDDED_LLAMA_CPP_PYTHON_VERSION = "0.2.84" \ No newline at end of file diff --git a/custom_components/llama_conversation/manifest.json b/custom_components/llama_conversation/manifest.json index d5c6da9..45104ea 100644 --- a/custom_components/llama_conversation/manifest.json +++ b/custom_components/llama_conversation/manifest.json @@ -1,7 +1,7 @@ { "domain": "llama_conversation", "name": "Local LLM Conversation", - "version": "0.3.3", + "version": "0.3.4", "codeowners": ["@acon96"], "config_flow": true, "dependencies": ["conversation"],