Update seed2_tokenizer.py
Browse files- seed2_tokenizer.py +2 -2
seed2_tokenizer.py
CHANGED
|
@@ -42,7 +42,7 @@ DIFFUSION_NAME = 'stabilityai/stable-diffusion-2-1-unclip'
|
|
| 42 |
# from qformer.qformer_quantizer import Blip2QformerQuantizer
|
| 43 |
# from diffusers import StableUnCLIPImg2ImgPipeline
|
| 44 |
|
| 45 |
-
from
|
| 46 |
|
| 47 |
import logging
|
| 48 |
|
|
@@ -2182,7 +2182,7 @@ class Seed2Tokenizer(PreTrainedModel):
|
|
| 2182 |
image_torch = self.processor(image_pil)
|
| 2183 |
|
| 2184 |
image_torch = image_torch.to(self.device)
|
| 2185 |
-
return self.encode(
|
| 2186 |
|
| 2187 |
if __name__ == "__main__":
|
| 2188 |
tokenizer = Seed2Tokenizer.from_pretrained("ontocord/seed2")
|
|
|
|
| 42 |
# from qformer.qformer_quantizer import Blip2QformerQuantizer
|
| 43 |
# from diffusers import StableUnCLIPImg2ImgPipeline
|
| 44 |
|
| 45 |
+
from diffusers import StableUnCLIPImg2ImgPipeline
|
| 46 |
|
| 47 |
import logging
|
| 48 |
|
|
|
|
| 2182 |
image_torch = self.processor(image_pil)
|
| 2183 |
|
| 2184 |
image_torch = image_torch.to(self.device)
|
| 2185 |
+
return self.encode(image_torch, visual_encoder)
|
| 2186 |
|
| 2187 |
if __name__ == "__main__":
|
| 2188 |
tokenizer = Seed2Tokenizer.from_pretrained("ontocord/seed2")
|