fix conversation: llama generates token not prob now (#3120)

This commit is contained in:
chenyu
2024-01-14 13:10:01 -05:00
committed by GitHub
parent 152ef7fc79
commit 79f4627fbc

View File

@@ -79,8 +79,7 @@ def llama_generate(
outputted = llama.tokenizer.decode(toks)
init_length = len(outputted)
for _ in range(max_tokens):
probs_np = llama.model(Tensor([toks[start_pos:]]), start_pos, temperature).numpy()
token = int(np.random.choice(len(probs_np), p=probs_np))
token = llama.model(Tensor([toks[start_pos:]]), start_pos, temperature).item()
start_pos = len(toks)
toks.append(token)