diff --git a/README.md b/README.md index 0f58bcc..a69c1e1 100644 --- a/README.md +++ b/README.md @@ -126,6 +126,7 @@ In order to facilitate running the project entirely on the system where Home Ass ## Version History | Version | Description | |---------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| +| v0.2.14 | Fix llama.cpp wheels + AVX detection | | v0.2.13 | Add support for Llama 3, build llama.cpp wheels that are compatible with non-AVX systems, fix an error with exposing script entities, fix multiple small Ollama backend issues, and add basic multi-language support | | v0.2.12 | Fix cover ICL examples, allow setting number of ICL examples, add min P and typical P sampler options, recommend models during setup, add JSON mode for Ollama backend, fix missing default options | | v0.2.11 | Add prompt caching, expose llama.cpp runtime settings, build llama-cpp-python wheels using GitHub actions, and install wheels directly from GitHub | diff --git a/addon/Dockerfile b/addon/Dockerfile index 1023514..b75756f 100644 --- a/addon/Dockerfile +++ b/addon/Dockerfile @@ -23,7 +23,7 @@ RUN \ python3-venv \ python3-pip \ \ - && git clone https://github.com/oobabooga/text-generation-webui.git ${APP_DIR} --branch snapshot-2024-04-14 \ + && git clone https://github.com/oobabooga/text-generation-webui.git ${APP_DIR} --branch snapshot-2024-04-28 \ && python3 -m pip install torch torchvision torchaudio py-cpuinfo==9.0.0 \ && python3 -m pip install -r ${APP_DIR}/requirements_cpu_only_noavx2.txt llama-cpp-python \ && apt-get purge -y --auto-remove \ diff --git a/addon/config.yaml b/addon/config.yaml index 310ce76..b33a174 100644 --- a/addon/config.yaml +++ b/addon/config.yaml @@ -1,6 +1,6 @@ --- name: oobabooga-text-generation-webui -version: 2024.04.14 +version: 2024.04.28 slug: text-generation-webui description: "A tool for running Large Language Models" url: "https://github.com/oobabooga/text-generation-webui" diff --git a/custom_components/llama_conversation/const.py b/custom_components/llama_conversation/const.py index 7310260..e235015 100644 --- a/custom_components/llama_conversation/const.py +++ b/custom_components/llama_conversation/const.py @@ -271,5 +271,5 @@ OPTIONS_OVERRIDES = { } } -INTEGRATION_VERSION = "0.2.13" -EMBEDDED_LLAMA_CPP_PYTHON_VERSION = "0.2.64" \ No newline at end of file +INTEGRATION_VERSION = "0.2.14" +EMBEDDED_LLAMA_CPP_PYTHON_VERSION = "0.2.69" \ No newline at end of file