cyd0806 commited on
Commit
54b90c1
·
verified ·
1 Parent(s): 5091451

Upload src/text_encoder.py with huggingface_hub

Browse files
Files changed (1) hide show
  1. src/text_encoder.py +118 -0
src/text_encoder.py ADDED
@@ -0,0 +1,118 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import ipdb
2
+ import torch
3
+
4
+ def _encode_prompt_with_t5(
5
+ text_encoder,
6
+ tokenizer,
7
+ max_sequence_length=512,
8
+ prompt=None,
9
+ num_images_per_prompt=1,
10
+ device=None,
11
+ text_input_ids=None,
12
+ ):
13
+ prompt = [prompt] if isinstance(prompt, str) else prompt
14
+ batch_size = len(prompt)
15
+
16
+ if tokenizer is not None:
17
+ text_inputs = tokenizer(
18
+ prompt,
19
+ padding="max_length",
20
+ max_length=max_sequence_length,
21
+ truncation=True,
22
+ return_length=False,
23
+ return_overflowing_tokens=False,
24
+ return_tensors="pt",
25
+ )
26
+ text_input_ids = text_inputs.input_ids
27
+ else:
28
+ if text_input_ids is None:
29
+ raise ValueError("text_input_ids must be provided when the tokenizer is not specified")
30
+
31
+ prompt_embeds = text_encoder(text_input_ids.to(device))[0]
32
+
33
+ dtype = text_encoder.dtype
34
+ prompt_embeds = prompt_embeds.to(dtype=dtype, device=device)
35
+
36
+ _, seq_len, _ = prompt_embeds.shape
37
+
38
+ # duplicate text embeddings and attention mask for each generation per prompt, using mps friendly method
39
+ prompt_embeds = prompt_embeds.repeat(1, num_images_per_prompt, 1)
40
+ prompt_embeds = prompt_embeds.view(batch_size * num_images_per_prompt, seq_len, -1)
41
+
42
+ return prompt_embeds
43
+
44
+
45
+ def _encode_prompt_with_clip(
46
+ text_encoder,
47
+ tokenizer,
48
+ prompt: str,
49
+ device=None,
50
+ text_input_ids=None,
51
+ num_images_per_prompt: int = 1,
52
+ ):
53
+ prompt = [prompt] if isinstance(prompt, str) else prompt
54
+ batch_size = len(prompt)
55
+
56
+ if tokenizer is not None:
57
+ text_inputs = tokenizer(
58
+ prompt,
59
+ padding="max_length",
60
+ max_length=77,
61
+ truncation=True,
62
+ return_overflowing_tokens=False,
63
+ return_length=False,
64
+ return_tensors="pt",
65
+ )
66
+
67
+ text_input_ids = text_inputs.input_ids
68
+ else:
69
+ if text_input_ids is None:
70
+ raise ValueError("text_input_ids must be provided when the tokenizer is not specified")
71
+
72
+ prompt_embeds = text_encoder(text_input_ids.to(device), output_hidden_states=False)
73
+
74
+ # Use pooled output of CLIPTextModel
75
+ prompt_embeds = prompt_embeds.pooler_output
76
+ prompt_embeds = prompt_embeds.to(dtype=text_encoder.dtype, device=device)
77
+
78
+ # duplicate text embeddings for each generation per prompt, using mps friendly method
79
+ prompt_embeds = prompt_embeds.repeat(1, num_images_per_prompt, 1)
80
+ prompt_embeds = prompt_embeds.view(batch_size * num_images_per_prompt, -1)
81
+
82
+ return prompt_embeds
83
+
84
+
85
+ def encode_prompt(
86
+ text_encoders,
87
+ tokenizers,
88
+ prompt: str,
89
+ max_sequence_length,
90
+ device=None,
91
+ num_images_per_prompt: int = 1,
92
+ text_input_ids_list=None,
93
+ ):
94
+ prompt = [prompt] if isinstance(prompt, str) else prompt
95
+ dtype = text_encoders[0].dtype
96
+
97
+ pooled_prompt_embeds = _encode_prompt_with_clip(
98
+ text_encoder=text_encoders[0],
99
+ tokenizer=tokenizers[0],
100
+ prompt=prompt,
101
+ device=device if device is not None else text_encoders[0].device,
102
+ num_images_per_prompt=num_images_per_prompt,
103
+ text_input_ids=text_input_ids_list[0] if text_input_ids_list else None,
104
+ )
105
+
106
+ prompt_embeds = _encode_prompt_with_t5(
107
+ text_encoder=text_encoders[1],
108
+ tokenizer=tokenizers[1],
109
+ max_sequence_length=max_sequence_length,
110
+ prompt=prompt,
111
+ num_images_per_prompt=num_images_per_prompt,
112
+ device=device if device is not None else text_encoders[1].device,
113
+ text_input_ids=text_input_ids_list[1] if text_input_ids_list else None,
114
+ )
115
+
116
+ text_ids = torch.zeros(prompt_embeds.shape[1], 3).to(device=device, dtype=dtype)
117
+
118
+ return prompt_embeds, pooled_prompt_embeds, text_ids