Release v0.2.14

This commit is contained in:
Alex O'Connell
2024-05-02 21:53:14 -04:00
parent 1cbba9e0d0
commit 875547d2e2
4 changed files with 5 additions and 4 deletions

View File

@@ -126,6 +126,7 @@ In order to facilitate running the project entirely on the system where Home Ass
## Version History
| Version | Description |
|---------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| v0.2.14 | Fix llama.cpp wheels + AVX detection |
| v0.2.13 | Add support for Llama 3, build llama.cpp wheels that are compatible with non-AVX systems, fix an error with exposing script entities, fix multiple small Ollama backend issues, and add basic multi-language support |
| v0.2.12 | Fix cover ICL examples, allow setting number of ICL examples, add min P and typical P sampler options, recommend models during setup, add JSON mode for Ollama backend, fix missing default options |
| v0.2.11 | Add prompt caching, expose llama.cpp runtime settings, build llama-cpp-python wheels using GitHub actions, and install wheels directly from GitHub |

View File

@@ -23,7 +23,7 @@ RUN \
python3-venv \
python3-pip \
\
&& git clone https://github.com/oobabooga/text-generation-webui.git ${APP_DIR} --branch snapshot-2024-04-14 \
&& git clone https://github.com/oobabooga/text-generation-webui.git ${APP_DIR} --branch snapshot-2024-04-28 \
&& python3 -m pip install torch torchvision torchaudio py-cpuinfo==9.0.0 \
&& python3 -m pip install -r ${APP_DIR}/requirements_cpu_only_noavx2.txt llama-cpp-python \
&& apt-get purge -y --auto-remove \

View File

@@ -1,6 +1,6 @@
---
name: oobabooga-text-generation-webui
version: 2024.04.14
version: 2024.04.28
slug: text-generation-webui
description: "A tool for running Large Language Models"
url: "https://github.com/oobabooga/text-generation-webui"

View File

@@ -271,5 +271,5 @@ OPTIONS_OVERRIDES = {
}
}
INTEGRATION_VERSION = "0.2.13"
EMBEDDED_LLAMA_CPP_PYTHON_VERSION = "0.2.64"
INTEGRATION_VERSION = "0.2.14"
EMBEDDED_LLAMA_CPP_PYTHON_VERSION = "0.2.69"