mirror of
https://github.com/tinygrad/tinygrad.git
synced 2026-01-08 14:43:57 -05:00
fix conversation: llama generates token not prob now (#3120)
This commit is contained in:
@@ -79,8 +79,7 @@ def llama_generate(
|
||||
outputted = llama.tokenizer.decode(toks)
|
||||
init_length = len(outputted)
|
||||
for _ in range(max_tokens):
|
||||
probs_np = llama.model(Tensor([toks[start_pos:]]), start_pos, temperature).numpy()
|
||||
token = int(np.random.choice(len(probs_np), p=probs_np))
|
||||
token = llama.model(Tensor([toks[start_pos:]]), start_pos, temperature).item()
|
||||
start_pos = len(toks)
|
||||
toks.append(token)
|
||||
|
||||
|
||||
Reference in New Issue
Block a user