diff --git a/README.md b/README.md index 40b5865..0f20122 100644 --- a/README.md +++ b/README.md @@ -136,6 +136,7 @@ In order to facilitate running the project entirely on the system where Home Ass ## Version History | Version | Description | |---------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| +| v0.3.4 | Update bundled llama-cpp-python to support new models, various bug fixes | | v0.3.3 | Improvements to the Generic OpenAI Backend, improved area handling, fix issue using RGB colors, remove EOS token from responses, replace requests dependency with aiohttp included with Home Assistant | | v0.3.2 | Fix for exposed script entities causing errors, fix missing GBNF error, trim whitespace from model output | | v0.3.1 | Adds basic area support in prompting, Fix for broken requirements, fix for issue with formatted tools, fix custom API not registering on startup properly | diff --git a/custom_components/llama_conversation/const.py b/custom_components/llama_conversation/const.py index d9d011c..e8a625d 100644 --- a/custom_components/llama_conversation/const.py +++ b/custom_components/llama_conversation/const.py @@ -340,5 +340,5 @@ OPTIONS_OVERRIDES = { }, } -INTEGRATION_VERSION = "0.3.3" +INTEGRATION_VERSION = "0.3.4" EMBEDDED_LLAMA_CPP_PYTHON_VERSION = "0.2.84" \ No newline at end of file diff --git a/custom_components/llama_conversation/manifest.json b/custom_components/llama_conversation/manifest.json index d5c6da9..45104ea 100644 --- a/custom_components/llama_conversation/manifest.json +++ b/custom_components/llama_conversation/manifest.json @@ -1,7 +1,7 @@ { "domain": "llama_conversation", "name": "Local LLM Conversation", - "version": "0.3.3", + "version": "0.3.4", "codeowners": ["@acon96"], "config_flow": true, "dependencies": ["conversation"],