WhiteAiZ's picture
Upload 1592 files
25fad07 verified
import torch
from modules.sd_hijack_clip import FrozenCLIPEmbedderWithCustomWords
from ldm_patched.modules import model_management
from modules import sd_models
from modules.shared import opts
class SD3CLIP(torch.nn.Module):
def __init__(self, clip_components, embedding_directory=None):
super().__init__()
self.clip_l, self.clip_g, self.t5xxl = clip_components
self.tokenizer = SD3Tokenizer(embedding_directory)
self.patcher = model_management.ModelPatcher(self)
def encode_with_transformers(self, tokens):
move_clip_to_gpu()
# Process tokens for each component
z_l, pooled_l = self.clip_l(tokens['l'])
z_g, pooled_g = self.clip_g(tokens['g'])
z_t5, _ = self.t5xxl(tokens['t5xxl'])
# Combine outputs
z = torch.cat([z_l, z_g, z_t5], dim=-2)
pooled = torch.cat([pooled_l, pooled_g], dim=-1)
return z, pooled
def encode_from_tokens(self, tokens, return_pooled=False):
z, pooled = self.encode_with_transformers(tokens)
if return_pooled:
return z, pooled
return z
class SD3Tokenizer:
def __init__(self, embedding_directory=None):
self.clip_l = self.clip_g.clip_l.tokenizer
self.clip_g = self.clip_g.clip_g.tokenizer
self.t5xxl = self.t5xxl.tokenizer
def tokenize(self, text):
return {
"l": self.clip_l.tokenize(text),
"g": self.clip_g.tokenize(text),
"t5xxl": self.t5xxl.tokenize(text)
}
def move_clip_to_gpu():
if sd_models.model_data.sd_model is None:
print('Error: CLIP called before SD is loaded!')
return
model_management.load_model_gpu(sd_models.model_data.sd_model.forge_objects.clip.patcher)