diff --git a/examples/gpt2.py b/examples/gpt2.py index ff4077ea6b..4c628c8fe6 100644 --- a/examples/gpt2.py +++ b/examples/gpt2.py @@ -1,11 +1,8 @@ #!/usr/bin/env python3 -import os +import os, argparse from typing import Optional, Union -import argparse -import numpy as np import tiktoken -from tinygrad import Tensor, TinyJit, Device, GlobalCounters, Variable -from tinygrad.dtype import dtypes +from tinygrad import Tensor, TinyJit, Device, GlobalCounters, Variable, dtypes from tinygrad.ops import UOp from tinygrad.helpers import Timing, DEBUG, JIT, getenv, fetch, colored, trange from tinygrad.nn import Embedding, Linear, LayerNorm @@ -218,7 +215,6 @@ if __name__ == "__main__": if args.seed is not None: Tensor.manual_seed(args.seed) - np.random.seed(args.seed) print(f"using {args.model_size}") gpt2 = GPT2.build_gguf(args.model_size) if args.model_size.startswith("gpt2_gguf_") else GPT2.build(args.model_size) diff --git a/examples/llama.py b/examples/llama.py index f07b0f6539..67ad7ec3a0 100755 --- a/examples/llama.py +++ b/examples/llama.py @@ -6,8 +6,6 @@ from pathlib import Path from typing import List, Optional import argparse, json -import numpy as np -np.set_printoptions(linewidth=200) from tinygrad import Tensor, Device, GlobalCounters, nn from tinygrad.helpers import Context, Timing, Profiling, DEBUG, JIT, getenv, colored from tinygrad.nn.state import safe_load, torch_load, load_state_dict, get_parameters @@ -250,6 +248,8 @@ class LLaMa: self.tokenizer = tokenizer def greedy_until(self, prompt:str, until, max_length, temperature): + # only used in old eval script + import numpy as np toks = [self.tokenizer.bos_id()] + self.tokenizer.encode(prompt) start_pos = 0 for i in range(max_length):