diff --git a/examples/compile_efficientnet.py b/examples/compile_efficientnet.py index 85d2232723..7817298315 100644 --- a/examples/compile_efficientnet.py +++ b/examples/compile_efficientnet.py @@ -34,7 +34,7 @@ if __name__ == "__main__": model = EfficientNet(0) model.load_from_pretrained() - from extra.jit import TinyJit + from tinygrad.jit import TinyJit @TinyJit def run(x): return model.forward(x).realize() diff --git a/examples/compile_tensorflow.py b/examples/compile_tensorflow.py index aef61bbf80..43e5685b29 100644 --- a/examples/compile_tensorflow.py +++ b/examples/compile_tensorflow.py @@ -29,7 +29,7 @@ def create_onnx_model(keras_model): def compile_onnx_model(onnx_model): run_onnx = get_run_onnx(onnx_model) - from extra.jit import TinyJit + from tinygrad.jit import TinyJit @TinyJit def run(x): return run_onnx({"x": x}, debug=False)['predictions'].realize() diff --git a/examples/efficientnet.py b/examples/efficientnet.py index 0ffa05c7fb..af6bf8acdc 100644 --- a/examples/efficientnet.py +++ b/examples/efficientnet.py @@ -12,7 +12,7 @@ from PIL import Image from tinygrad.tensor import Tensor from tinygrad.helpers import getenv from extra.utils import fetch -from extra.jit import TinyJit +from tinygrad.jit import TinyJit from models.efficientnet import EfficientNet np.set_printoptions(suppress=True) diff --git a/examples/hlb_cifar10.py b/examples/hlb_cifar10.py index 10d8a07a63..ee76ad8203 100644 --- a/examples/hlb_cifar10.py +++ b/examples/hlb_cifar10.py @@ -49,7 +49,7 @@ class SpeedyResNet: # note, pytorch just uses https://pytorch.org/docs/stable/generated/torch.nn.CrossEntropyLoss.html instead of log_softmax def __call__(self, x): return x.sequential(self.net).log_softmax() -from extra.jit import TinyJit +from tinygrad.jit import TinyJit @TinyJit def train_step_jitted(model, optimizer, X, Y): out = model(X) diff --git a/openpilot/compile.py b/openpilot/compile.py index 91a235a0c6..5161dc9b1c 100644 --- a/openpilot/compile.py +++ b/openpilot/compile.py @@ -33,7 +33,7 @@ def get_random_input_tensors(input_shapes): np_inputs = {k:v.realize().numpy() for k,v in inputs.items()} return inputs, np_inputs -from extra.jit import TinyJit +from tinygrad.jit import TinyJit @TinyJit def model_exec(run_onnx, using_graph, **inputs): diff --git a/test/test_jit.py b/test/test_jit.py index 04d25a3a47..dcbdf60cfd 100644 --- a/test/test_jit.py +++ b/test/test_jit.py @@ -2,7 +2,7 @@ import unittest import numpy as np from tinygrad.tensor import Tensor, Device -from extra.jit import TinyJit +from tinygrad.jit import TinyJit @unittest.skipUnless(Device.DEFAULT == "GPU", "JIT is only for GPU") class TestJit(unittest.TestCase): diff --git a/test/test_speed_v_torch.py b/test/test_speed_v_torch.py index 994109e585..eca6a4ae29 100644 --- a/test/test_speed_v_torch.py +++ b/test/test_speed_v_torch.py @@ -11,7 +11,7 @@ from tinygrad.ops import GlobalCounters from tinygrad.tensor import Tensor from tinygrad.nn import Conv2d from tinygrad.helpers import colored, getenv, DEBUG -from extra.jit import TinyJit +from tinygrad.jit import TinyJit METAL = getenv("METAL") try: from tinygrad.runtime.opencl import CL diff --git a/extra/jit.py b/tinygrad/jit.py similarity index 97% rename from extra/jit.py rename to tinygrad/jit.py index ee3d6355eb..94c86d8219 100644 --- a/extra/jit.py +++ b/tinygrad/jit.py @@ -2,7 +2,6 @@ from typing import Callable, List, Tuple import itertools from tinygrad.lazy import Device from tinygrad.tensor import Tensor -from tinygrad.helpers import DEBUG from tinygrad.ops import GlobalCounters class TinyJit: