fix outputs

This commit is contained in:
Alex O'Connell
2024-01-13 15:37:28 -05:00
parent e479b5ffac
commit 9bfd3bc49d
3 changed files with 4 additions and 4 deletions

View File

@@ -230,7 +230,7 @@ class LLaMAAgent(conversation.AbstractConversationAgent):
self.history[conversation_id] = prompt
exposed_entities = list(self._async_get_exposed_entities()[0].keys())
pattern = re.compile(r"```homeassistant\n([\S\n]*)```")
pattern = re.compile(r"```homeassistant\n([\S \t\n]*)```")
to_say = pattern.sub("", response).strip()
for block in pattern.findall(response.strip()):

View File

@@ -1,6 +1,6 @@
root ::= ws tosay ws functioncalls?
root ::= (ws tosay)+ ws functioncalls?
tosay ::= [0-9a-zA-Z ]*
tosay ::= [0-9a-zA-Z .#%]*
functioncalls ::=
"```homeassistant\n" (object ws)* "```"

View File

@@ -11,4 +11,4 @@ if [[ ! -d "./models/$MODEL_NAME" ]]; then
fi
PROMPT=$(cat $PROMPT_SRC)
$LLAMA_CPP/build/bin/main --model "./models/$MODEL_NAME/$MODEL_NAME.$QUANT_TYPE.gguf" --temp 0.1 --ctx-size 2048 --prompt "$PROMPT" --grammar-file ./custom_components/llama_conversation/output.gbnf
$LLAMA_CPP/build/bin/main --model "./models/$MODEL_NAME/$MODEL_NAME.$QUANT_TYPE.gguf" --temp 0.1 --ctx-size 2048 --prompt "$PROMPT" --grammar-file ./custom_components/llama_conversation/output.gbnf