mirror of
https://github.com/acon96/home-llm.git
synced 2026-01-09 13:48:05 -05:00
Release v0.2.14
This commit is contained in:
@@ -126,6 +126,7 @@ In order to facilitate running the project entirely on the system where Home Ass
|
||||
## Version History
|
||||
| Version | Description |
|
||||
|---------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
|
||||
| v0.2.14 | Fix llama.cpp wheels + AVX detection |
|
||||
| v0.2.13 | Add support for Llama 3, build llama.cpp wheels that are compatible with non-AVX systems, fix an error with exposing script entities, fix multiple small Ollama backend issues, and add basic multi-language support |
|
||||
| v0.2.12 | Fix cover ICL examples, allow setting number of ICL examples, add min P and typical P sampler options, recommend models during setup, add JSON mode for Ollama backend, fix missing default options |
|
||||
| v0.2.11 | Add prompt caching, expose llama.cpp runtime settings, build llama-cpp-python wheels using GitHub actions, and install wheels directly from GitHub |
|
||||
|
||||
@@ -23,7 +23,7 @@ RUN \
|
||||
python3-venv \
|
||||
python3-pip \
|
||||
\
|
||||
&& git clone https://github.com/oobabooga/text-generation-webui.git ${APP_DIR} --branch snapshot-2024-04-14 \
|
||||
&& git clone https://github.com/oobabooga/text-generation-webui.git ${APP_DIR} --branch snapshot-2024-04-28 \
|
||||
&& python3 -m pip install torch torchvision torchaudio py-cpuinfo==9.0.0 \
|
||||
&& python3 -m pip install -r ${APP_DIR}/requirements_cpu_only_noavx2.txt llama-cpp-python \
|
||||
&& apt-get purge -y --auto-remove \
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
---
|
||||
name: oobabooga-text-generation-webui
|
||||
version: 2024.04.14
|
||||
version: 2024.04.28
|
||||
slug: text-generation-webui
|
||||
description: "A tool for running Large Language Models"
|
||||
url: "https://github.com/oobabooga/text-generation-webui"
|
||||
|
||||
@@ -271,5 +271,5 @@ OPTIONS_OVERRIDES = {
|
||||
}
|
||||
}
|
||||
|
||||
INTEGRATION_VERSION = "0.2.13"
|
||||
EMBEDDED_LLAMA_CPP_PYTHON_VERSION = "0.2.64"
|
||||
INTEGRATION_VERSION = "0.2.14"
|
||||
EMBEDDED_LLAMA_CPP_PYTHON_VERSION = "0.2.69"
|
||||
Reference in New Issue
Block a user