# https://arxiv.org/pdf/2112.10752.pdf # https://github.com/ekagra-ranjan/huggingface-blog/blob/main/stable_diffusion.md import os import numpy as np import traceback from collections import namedtuple from extra.utils import fake_torch_load_zipped, get_child from tinygrad.nn import Conv2d from tinygrad.tensor import Tensor # TODO: rename to GroupNorm and put in nn.py class Normalize: def __init__(self, in_channels, num_groups=32): self.weight = Tensor.empty(in_channels) self.bias = Tensor.empty(in_channels) self.num_groups = num_groups def __call__(self, x): # reshape for layernorm to work as group norm # subtract mean and divide stddev x = x.reshape(x.shape[0], self.num_groups, -1).layernorm().reshape(x.shape) # elementwise_affine on channels if len(x.shape) == 4: # HACK for channels in conv return (x * self.weight.reshape(1, -1, 1, 1)) + self.bias.reshape(1, -1, 1, 1) else: return x.linear(self.weight, self.bias) class AttnBlock: def __init__(self, in_channels): self.norm = Normalize(in_channels) self.q = Conv2d(in_channels, in_channels, 1) self.k = Conv2d(in_channels, in_channels, 1) self.v = Conv2d(in_channels, in_channels, 1) self.proj_out = Conv2d(in_channels, in_channels, 1) # copied from AttnBlock in ldm repo def __call__(self, x): h_ = self.norm(x) q,k,v = self.q(h_), self.k(h_), self.v(h_) # compute attention b,c,h,w = q.shape q = q.reshape(b,c,h*w) q = q.permute(0,2,1) # b,hw,c k = k.reshape(b,c,h*w) # b,c,hw w_ = q @ k w_ = w_ * (c**(-0.5)) w_ = w_.softmax() # attend to values v = v.reshape(b,c,h*w) w_ = w_.permute(0,2,1) h_ = v @ w_ h_ = h_.reshape(b,c,h,w) return x + self.proj_out(h_) class ResnetBlock: def __init__(self, in_channels, out_channels=None): self.norm1 = Normalize(in_channels) self.conv1 = Conv2d(in_channels, out_channels, 3, padding=1) self.norm2 = Normalize(out_channels) self.conv2 = Conv2d(out_channels, out_channels, 3, padding=1) self.nin_shortcut = Conv2d(in_channels, out_channels, 1) if in_channels != out_channels else lambda x: x def __call__(self, x): h = self.conv1(self.norm1(x).swish()) h = self.conv2(self.norm2(h).swish()) return self.nin_shortcut(x) + h class Mid: def __init__(self, block_in): self.block_1 = ResnetBlock(block_in, block_in) self.attn_1 = AttnBlock(block_in) self.block_2 = ResnetBlock(block_in, block_in) def __call__(self, x): return x.sequential([self.block_1, self.attn_1, self.block_2]) class Decoder: def __init__(self): sz = [(128, 256), (256, 512), (512, 512), (512, 512)] self.conv_in = Conv2d(4,512,3, padding=1) self.mid = Mid(512) arr = [] for i,s in enumerate(sz): arr.append({"block": [ResnetBlock(s[1], s[0]), ResnetBlock(s[0], s[0]), ResnetBlock(s[0], s[0])]}) if i != 0: arr[-1]['upsample'] = {"conv": Conv2d(s[0], s[0], 3, padding=1)} self.up = arr self.norm_out = Normalize(128) self.conv_out = Conv2d(128, 3, 3, padding=1) def __call__(self, x): x = self.conv_in(x) x = self.mid(x) for l in self.up[::-1]: print("decode", x.shape) for b in l['block']: x = b(x) if 'upsample' in l: # https://pytorch.org/docs/stable/generated/torch.nn.functional.interpolate.html ? bs,c,py,px = x.shape x = x.reshape(bs, c, py, 1, px, 1).expand(bs, c, py, 2, px, 2).reshape(bs, c, py*2, px*2) x = l['upsample']['conv'](x) return self.conv_out(self.norm_out(x).swish()) class Encoder: def __init__(self): sz = [(128, 128), (128, 256), (256, 512), (512, 512)] self.conv_in = Conv2d(3,128,3, padding=1) arr = [] for i,s in enumerate(sz): arr.append({"block": [ResnetBlock(s[0], s[1]), ResnetBlock(s[1], s[1])]}) if i != 3: arr[-1]['downsample'] = {"conv": Conv2d(s[1], s[1], 3, stride=2, padding=(0,1,0,1))} self.down = arr self.mid = Mid(512) self.norm_out = Normalize(512) self.conv_out = Conv2d(512, 8, 3, padding=1) def __call__(self, x): x = self.conv_in(x) for l in self.down: print("encode", x.shape) for b in l['block']: x = b(x) if 'downsample' in l: x = l['downsample']['conv'](x) x = self.mid(x) return self.conv_out(self.norm_out(x).swish()) class AutoencoderKL: def __init__(self): self.encoder = Encoder() self.decoder = Decoder() self.quant_conv = Conv2d(8, 8, 1) self.post_quant_conv = Conv2d(4, 4, 1) def __call__(self, x): latent = self.encoder(x) latent = self.quant_conv(latent) latent = latent[:, 0:4] # only the means print("latent", latent.shape) latent = self.post_quant_conv(latent) return self.decoder(latent) class Linear: def __init__(self, in_features, out_features, bias=True): self.weight = Tensor.empty(out_features, in_features) self.bias = Tensor.empty(out_features) if bias else None def __call__(self, x): #print(x.shape, self.weight.shape, self.bias.shape) return x.linear(self.weight.transpose(), self.bias) # not to be confused with ResnetBlock class ResBlock: def __init__(self, channels, emb_channels, out_channels): self.in_layers = [ Normalize(channels), Tensor.silu, Conv2d(channels, out_channels, 3, padding=1) ] self.emb_layers = [ Tensor.silu, Linear(emb_channels, out_channels) ] self.out_layers = [ Normalize(out_channels), Tensor.silu, lambda x: x, Conv2d(out_channels, out_channels, 3, padding=1) ] self.skip_connection = Conv2d(channels, out_channels, 1) if channels != out_channels else lambda x: x def __call__(self, x, emb): h = x.sequential(self.in_layers) emb_out = emb.sequential(self.emb_layers) h = h + emb_out h = h.sequential(self.out_layers) return self.skip_connection(x) + h class CrossAttention: def __init__(self, query_dim, context_dim, n_heads, d_head): self.to_q = Linear(query_dim, n_heads*d_head, bias=False) self.to_k = Linear(context_dim, n_heads*d_head, bias=False) self.to_v = Linear(context_dim, n_heads*d_head, bias=False) self.to_out = [Linear(n_heads*d_head, query_dim)] # TODO: this is probably very wrong def __call__(self, x, context=None): context = x if context is None else context q,k,v = self.to_q(x), self.to_k(context), self.to_v(context) # compute attention b,hw,c = q.shape print("cross attention", q.shape, k.shape, v.shape) k = k.permute(0,2,1) # b,c,hw w_ = q @ k w_ = w_ * (c**(-0.5)) w_ = w_.softmax() # attend to values # TODO: ugh this is probably wrong #print(v.shape, w_.shape) h_ = w_ @ v #print(h_.shape) return h_.sequential(self.to_out) class GEGLU: def __init__(self, dim_in, dim_out): self.proj = Linear(dim_in, dim_out * 2) self.dim_out = dim_out def __call__(self, x): x, gate = self.proj(x).chunk(2, dim=-1) return x * gate.gelu() class FeedForward: def __init__(self, dim, mult=4): self.net = [ GEGLU(dim, dim*mult), lambda x: x, Linear(dim*mult, dim) ] def __call__(self, x): return x.sequential(self.net) class BasicTransformerBlock: def __init__(self, dim, context_dim, n_heads, d_head): self.attn1 = CrossAttention(dim, dim, n_heads, d_head) self.ff = FeedForward(dim) self.attn2 = CrossAttention(dim, context_dim, n_heads, d_head) self.norm1 = Normalize(dim, num_groups=1) self.norm2 = Normalize(dim, num_groups=1) self.norm3 = Normalize(dim, num_groups=1) def __call__(self, x, context=None): x = self.attn1(self.norm1(x)) + x x = self.attn2(self.norm2(x), context=context) + x x = self.ff(self.norm3(x)) + x return x class SpatialTransformer: def __init__(self, channels, context_dim, n_heads, d_head): self.norm = Normalize(channels) self.proj_in = Conv2d(channels, n_heads * d_head, 1) self.transformer_blocks = [BasicTransformerBlock(channels, context_dim, n_heads, d_head)] self.proj_out = Conv2d(n_heads * d_head, channels, 1) def __call__(self, x, context=None): b, c, h, w = x.shape x_in = x x = self.norm(x) x = self.proj_in(x) x = x.reshape(b, c, h*w).permute(0,2,1) for block in self.transformer_blocks: x = block(x, context=context) x = x.permute(0,2,1).reshape(b, c, h, w) return self.proj_out(x) + x_in class Downsample: def __init__(self, channels): self.op = Conv2d(channels, channels, 3, stride=2, padding=(0,1,0,1)) def __call__(self, x): return self.op(x) class Upsample: def __init__(self, channels): self.conv = Conv2d(channels, channels, 3, padding=1) def __call__(self, x): bs,c,py,px = x.shape x = x.reshape(bs, c, py, 1, px, 1).expand(bs, c, py, 2, px, 2).reshape(bs, c, py*2, px*2) return self.conv(x) class UNetModel: def __init__(self): self.time_embed = [ Linear(320, 1280), Tensor.silu, Linear(1280, 1280), ] self.input_blocks = [ [Conv2d(4, 320, kernel_size=3, padding=1)], # TODO: my head sizes and counts are a guess [ResBlock(320, 1280, 320), SpatialTransformer(320, 768, 10, 32)], [ResBlock(320, 1280, 320), SpatialTransformer(320, 768, 10, 32)], [Downsample(320)], [ResBlock(320, 1280, 640), SpatialTransformer(640, 768, 10, 64)], [ResBlock(640, 1280, 640), SpatialTransformer(640, 768, 10, 64)], [Downsample(640)], [ResBlock(640, 1280, 1280), SpatialTransformer(1280, 768, 10, 128)], [ResBlock(1280, 1280, 1280), SpatialTransformer(1280, 768, 10, 128)], [Downsample(1280)], [ResBlock(1280, 1280, 1280)], [ResBlock(1280, 1280, 1280)] ] self.middle_block = [ ResBlock(1280, 1280, 1280), SpatialTransformer(1280, 768, 10, 128), ResBlock(1280, 1280, 1280) ] self.output_blocks = [ [ResBlock(2560, 1280, 1280)], [ResBlock(2560, 1280, 1280)], [ResBlock(2560, 1280, 1280), Upsample(1280)], [ResBlock(2560, 1280, 1280), SpatialTransformer(1280, 768, 10, 128)], [ResBlock(2560, 1280, 1280), SpatialTransformer(1280, 768, 10, 128)], [ResBlock(1920, 1280, 1280), SpatialTransformer(1280, 768, 10, 128), Upsample(1280)], [ResBlock(1920, 1280, 640), SpatialTransformer(640, 768, 10, 64)], # 6 [ResBlock(1280, 1280, 640), SpatialTransformer(640, 768, 10, 64)], [ResBlock(960, 1280, 640), SpatialTransformer(640, 768, 10, 64), Upsample(640)], [ResBlock(960, 1280, 320), SpatialTransformer(320, 768, 10, 32)], [ResBlock(640, 1280, 320), SpatialTransformer(320, 768, 10, 32)], [ResBlock(640, 1280, 320), SpatialTransformer(320, 768, 10, 32)], ] self.out = [ Normalize(320), Tensor.silu, Conv2d(320, 4, kernel_size=3, padding=1) ] def __call__(self, x, context=None): # TODO: real time embedding t_emb = Tensor.uniform(x.shape[0], 320) emb = t_emb.sequential(self.time_embed) def run(x, bb): if isinstance(bb, ResBlock): x = bb(x, emb) elif isinstance(bb, SpatialTransformer): x = bb(x, context) else: x = bb(x) return x saved_inputs = [] for i,b in enumerate(self.input_blocks): print("input block", i) for bb in b: x = run(x, bb) saved_inputs.append(x) for bb in self.middle_block: x = run(x, bb) for i,b in enumerate(self.output_blocks): print("output block", i) x = x.cat(saved_inputs.pop(), dim=1) for bb in b: x = run(x, bb) return x.sequential(self.out) class StableDiffusion: def __init__(self): self.model = namedtuple("DiffusionModel", ["diffusion_model"])(diffusion_model = UNetModel()) self.first_stage_model = AutoencoderKL() def __call__(self, x): context = Tensor.uniform(1, 77, 768) return self.model.diffusion_model(x, context) #return self.first_stage_model(x) # ** ldm.models.autoencoder.AutoencoderKL (done!) # 3x512x512 <--> 4x64x64 (16384) # decode torch.Size([1, 4, 64, 64]) torch.Size([1, 3, 512, 512]) # section 4.3 of paper # first_stage_model.encoder, first_stage_model.decoder # ** ldm.modules.diffusionmodules.openaimodel.UNetModel # this is what runs each time to sample. is this the LDM? # input: 4x64x64 # output: 4x64x64 # model.diffusion_model # it has attention? # ** ldm.modules.encoders.modules.FrozenCLIPEmbedder # cond_stage_model.transformer.text_model # this is sd-v1-4.ckpt #FILENAME = "/Users/kafka/fun/mps/stable-diffusion/models/ldm/stable-diffusion-v1/model.ckpt" FILENAME = "/home/kafka/model.ckpt" REAL = int(os.getenv("REAL", 0)) if __name__ == "__main__": Tensor.no_init = True # WTF!! no_grad brakes it #Tensor.no_grad = True model = StableDiffusion() # load in weights dat = fake_torch_load_zipped(open(FILENAME, "rb"), load_weights=REAL) for k,v in dat['state_dict'].items(): try: w = get_child(model, k) except (AttributeError, KeyError, IndexError): #traceback.print_exc() w = None print(f"{str(v.shape):30s}", w, k) if w is not None: assert w.shape == v.shape w.assign(v.astype(np.float32)) # load apple latent space nz = Tensor(np.load("datasets/stable_diffusion_apple.npy")) # run one pass of unet nz = model(nz) del model.model # clear unet nz = nz.detach() import gc gc.collect() import torch torch.cuda.empty_cache() """ print(out) print(out.numpy()) exit(0) if not REAL: exit(0) """ # load image #IMG = "/tmp/apple.png" #from PIL import Image #realimg = Tensor(np.array(Image.open(IMG))).permute((2,0,1)).reshape((1,3,512,512))*(1/255) #print(realimg.shape) #x = model(realimg) # load latent space x = model.first_stage_model.post_quant_conv(nz) x = model.first_stage_model.decoder(x) x = x.reshape(3,512,512).permute(1,2,0) dat = (x.detach().numpy().clip(0, 1)*255).astype(np.uint8) print(dat.shape) from PIL import Image im = Image.fromarray(dat) im.save("/tmp/rendered.png") # torch junk #IMG = "/Users/kafka/fun/mps/stable-diffusion/outputs/txt2img-samples/grid-0006.png" #from PIL import Image #realimg = Tensor(np.array(Image.open(IMG))).permute((2,0,1)).reshape((1,3,512,512))*(1/255) #print(img.shape) #x = model(img) #nz = np.random.randn(*nz.shape) * 100 # PYTHONPATH="$PWD:/Users/kafka/fun/mps/stable-diffusion" """ from ldm.models.autoencoder import AutoencoderKL import torch ckpt = torch.load(FILENAME) dat = ckpt['state_dict'] sd = {} for k in dat: if k.startswith("first_stage_model."): sd[k[len("first_stage_model."):]] = dat[k] print("loading", len(sd)) tmodel = AutoencoderKL( ddconfig = { "double_z": True, "z_channels": 4, "resolution": 256, "in_channels": 3, "out_ch": 3, "ch": 128, "ch_mult": [1,2,4,4], "num_res_blocks": 2, "attn_resolutions": [] }, lossconfig={"target": "torch.nn.Identity"}, embed_dim=4) tmodel.load_state_dict(sd, strict=True) nz = np.load("datasets/stable_diffusion_apple.npy") zmodel = model.first_stage_model x_torch = torch.tensor(nz) x_tiny = Tensor(nz) x_torch = tmodel.post_quant_conv(x_torch) x_tiny = zmodel.post_quant_conv(x_tiny) x_torch = tmodel.decoder.conv_in(x_torch) x_tiny = zmodel.decoder.conv_in(x_tiny) x_torch = tmodel.decoder.mid.block_1(x_torch, None) x_tiny = zmodel.decoder.mid['block_1'](x_tiny) """ """ x_torch = tmodel.decoder.mid.block_1.norm1(x_torch) x_tiny = zmodel.decoder.mid['block_1'].norm1(x_tiny) x_torch = x_torch * torch.sigmoid(x_torch) x_tiny = x_tiny.swish() print(zmodel.decoder.mid['block_1'].conv1.weight.shape) print(x_tiny.shape) x_torch = tmodel.decoder.mid.block_1.conv1(x_torch) x_tiny = zmodel.decoder.mid['block_1'].conv1(x_tiny) """ #print(tmodel.decoder.mid.block_1.conv1.weight) #print(zmodel.decoder.mid['block_1'].conv1.weight.numpy()) #print(abs(x_torch.detach().numpy() - x_tiny.numpy()).mean()) #print(x_torch.shape, x_tiny.shape) #exit(0) #exit(0) """ posterior = tmodel.encode(torch.tensor(realimg.numpy())) z = posterior.mode() print(z.shape) #exit(0) nz = z.detach().numpy() np.save("/tmp/apple.npy", nz) exit(0) """ #x, latent = tmodel(torch.tensor(realimg.numpy())) #x = tmodel.decode(torch.tensor(nz)) #x = x.reshape(3,512,512).permute(1,2,0) """ x = Tensor.randn(1,4,64,64) x = model.first_stage_model.post_quant_conv(x) x = model.first_stage_model.decoder(x) print(x.shape) x = x.reshape((3,512,512)).permute((1,2,0)) print(x.shape) if not REAL: exit(0) """ """ #dat = (x.detach().numpy()*256).astype(np.uint8) dat = (x.detach().numpy().clip(0, 1)*255).astype(np.uint8) print(dat.shape) from PIL import Image im = Image.fromarray(dat) im.save("/tmp/rendered.png") """