|
|
import torch |
|
|
from modules.sd_hijack_clip import FrozenCLIPEmbedderWithCustomWords |
|
|
from ldm_patched.modules import model_management |
|
|
from modules import sd_models |
|
|
from modules.shared import opts |
|
|
|
|
|
class SD3CLIP(torch.nn.Module): |
|
|
def __init__(self, clip_components, embedding_directory=None): |
|
|
super().__init__() |
|
|
self.clip_l, self.clip_g, self.t5xxl = clip_components |
|
|
self.tokenizer = SD3Tokenizer(embedding_directory) |
|
|
self.patcher = model_management.ModelPatcher(self) |
|
|
|
|
|
def encode_with_transformers(self, tokens): |
|
|
move_clip_to_gpu() |
|
|
|
|
|
|
|
|
z_l, pooled_l = self.clip_l(tokens['l']) |
|
|
z_g, pooled_g = self.clip_g(tokens['g']) |
|
|
z_t5, _ = self.t5xxl(tokens['t5xxl']) |
|
|
|
|
|
|
|
|
z = torch.cat([z_l, z_g, z_t5], dim=-2) |
|
|
pooled = torch.cat([pooled_l, pooled_g], dim=-1) |
|
|
|
|
|
return z, pooled |
|
|
|
|
|
def encode_from_tokens(self, tokens, return_pooled=False): |
|
|
z, pooled = self.encode_with_transformers(tokens) |
|
|
if return_pooled: |
|
|
return z, pooled |
|
|
return z |
|
|
|
|
|
class SD3Tokenizer: |
|
|
def __init__(self, embedding_directory=None): |
|
|
self.clip_l = self.clip_g.clip_l.tokenizer |
|
|
self.clip_g = self.clip_g.clip_g.tokenizer |
|
|
self.t5xxl = self.t5xxl.tokenizer |
|
|
|
|
|
def tokenize(self, text): |
|
|
return { |
|
|
"l": self.clip_l.tokenize(text), |
|
|
"g": self.clip_g.tokenize(text), |
|
|
"t5xxl": self.t5xxl.tokenize(text) |
|
|
} |
|
|
|
|
|
def move_clip_to_gpu(): |
|
|
if sd_models.model_data.sd_model is None: |
|
|
print('Error: CLIP called before SD is loaded!') |
|
|
return |
|
|
model_management.load_model_gpu(sd_models.model_data.sd_model.forge_objects.clip.patcher) |