mirror of
https://github.com/acon96/home-llm.git
synced 2026-01-09 21:58:00 -05:00
Release v0.3.4
This commit is contained in:
@@ -136,6 +136,7 @@ In order to facilitate running the project entirely on the system where Home Ass
|
|||||||
## Version History
|
## Version History
|
||||||
| Version | Description |
|
| Version | Description |
|
||||||
|---------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
|
|---------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
|
||||||
|
| v0.3.4 | Update bundled llama-cpp-python to support new models, various bug fixes |
|
||||||
| v0.3.3 | Improvements to the Generic OpenAI Backend, improved area handling, fix issue using RGB colors, remove EOS token from responses, replace requests dependency with aiohttp included with Home Assistant |
|
| v0.3.3 | Improvements to the Generic OpenAI Backend, improved area handling, fix issue using RGB colors, remove EOS token from responses, replace requests dependency with aiohttp included with Home Assistant |
|
||||||
| v0.3.2 | Fix for exposed script entities causing errors, fix missing GBNF error, trim whitespace from model output |
|
| v0.3.2 | Fix for exposed script entities causing errors, fix missing GBNF error, trim whitespace from model output |
|
||||||
| v0.3.1 | Adds basic area support in prompting, Fix for broken requirements, fix for issue with formatted tools, fix custom API not registering on startup properly |
|
| v0.3.1 | Adds basic area support in prompting, Fix for broken requirements, fix for issue with formatted tools, fix custom API not registering on startup properly |
|
||||||
|
|||||||
@@ -340,5 +340,5 @@ OPTIONS_OVERRIDES = {
|
|||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
INTEGRATION_VERSION = "0.3.3"
|
INTEGRATION_VERSION = "0.3.4"
|
||||||
EMBEDDED_LLAMA_CPP_PYTHON_VERSION = "0.2.84"
|
EMBEDDED_LLAMA_CPP_PYTHON_VERSION = "0.2.84"
|
||||||
@@ -1,7 +1,7 @@
|
|||||||
{
|
{
|
||||||
"domain": "llama_conversation",
|
"domain": "llama_conversation",
|
||||||
"name": "Local LLM Conversation",
|
"name": "Local LLM Conversation",
|
||||||
"version": "0.3.3",
|
"version": "0.3.4",
|
||||||
"codeowners": ["@acon96"],
|
"codeowners": ["@acon96"],
|
||||||
"config_flow": true,
|
"config_flow": true,
|
||||||
"dependencies": ["conversation"],
|
"dependencies": ["conversation"],
|
||||||
|
|||||||
Reference in New Issue
Block a user