From ebb81e8f11b3644a5a45dc472db834dcde480424 Mon Sep 17 00:00:00 2001 From: George Hotz Date: Fri, 5 Jan 2024 20:18:52 -0800 Subject: [PATCH] hotfix: st.size() -> st.size in llama --- examples/llama.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/examples/llama.py b/examples/llama.py index d4ba310fab..077a8e5603 100755 --- a/examples/llama.py +++ b/examples/llama.py @@ -370,7 +370,7 @@ After you are done speaking, output [EOS]. You are not Chad. TOKENIZER_PATH = (MODEL_PATH if MODEL_PATH.is_dir() else MODEL_PATH.parent) / "tokenizer.model" print(f"using LLaMA{LLAMA_SUFFIX}-{args.size} model") llama = LLaMa.build(MODEL_PATH, TOKENIZER_PATH, model_gen=args.gen, model_size=args.size, quantize=args.quantize) - param_count = sum(x.lazydata.st.size() for x in get_parameters(llama.model)) + param_count = sum(x.lazydata.st.size for x in get_parameters(llama.model)) if chatbot: # encode pre prompt