Update modeling_minicpmv.py
Browse files- modeling_minicpmv.py +3 -2
modeling_minicpmv.py
CHANGED
|
@@ -102,7 +102,7 @@ class MiniCPMV(MiniCPMVPreTrainedModel):
|
|
| 102 |
padding_value=0.0)
|
| 103 |
B, L, _ = all_pixel_values.shape
|
| 104 |
all_pixel_values = all_pixel_values.permute(0, 2, 1).reshape(B, 3, -1, L)
|
| 105 |
-
|
| 106 |
patch_attn_mask = torch.zeros((B, 1, max_patches), dtype=torch.bool, device=device)
|
| 107 |
for i in range(B):
|
| 108 |
patch_attn_mask[i, :tgt_sizes[i][0] * tgt_sizes[i][1]] = True
|
|
@@ -168,7 +168,8 @@ class MiniCPMV(MiniCPMVPreTrainedModel):
|
|
| 168 |
cur_vs_hs.view(-1, cur_vs_hs.shape[-1]))
|
| 169 |
elif self.training:
|
| 170 |
cur_vllm_emb += cur_vs_hs[0].mean() * 0
|
| 171 |
-
|
|
|
|
| 172 |
return vllm_embedding, vision_hidden_states
|
| 173 |
|
| 174 |
def forward(self, data, **kwargs):
|
|
|
|
| 102 |
padding_value=0.0)
|
| 103 |
B, L, _ = all_pixel_values.shape
|
| 104 |
all_pixel_values = all_pixel_values.permute(0, 2, 1).reshape(B, 3, -1, L)
|
| 105 |
+
print(B, "BATCH")
|
| 106 |
patch_attn_mask = torch.zeros((B, 1, max_patches), dtype=torch.bool, device=device)
|
| 107 |
for i in range(B):
|
| 108 |
patch_attn_mask[i, :tgt_sizes[i][0] * tgt_sizes[i][1]] = True
|
|
|
|
| 168 |
cur_vs_hs.view(-1, cur_vs_hs.shape[-1]))
|
| 169 |
elif self.training:
|
| 170 |
cur_vllm_emb += cur_vs_hs[0].mean() * 0
|
| 171 |
+
|
| 172 |
+
print(vllm_embedding.shape)
|
| 173 |
return vllm_embedding, vision_hidden_states
|
| 174 |
|
| 175 |
def forward(self, data, **kwargs):
|