Update modeling_minicpmv.py
Browse files- modeling_minicpmv.py +2 -2
modeling_minicpmv.py
CHANGED
|
@@ -273,7 +273,7 @@ class MiniCPMV(MiniCPMVPreTrainedModel):
|
|
| 273 |
else:
|
| 274 |
model_inputs["vision_hidden_states"] = vision_hidden_states
|
| 275 |
|
| 276 |
-
print(model_inputs)
|
| 277 |
(
|
| 278 |
input_embeds,
|
| 279 |
vision_hidden_states_dummy,
|
|
@@ -304,7 +304,7 @@ class MiniCPMV(MiniCPMVPreTrainedModel):
|
|
| 304 |
print(stuff.shape, "SHAPE")
|
| 305 |
batch = torch.cat(padded_tensors, dim=0)
|
| 306 |
print(batch.shape)
|
| 307 |
-
|
| 308 |
# output_ids = self._decode(input_embeds, tokenizer, **kwargs)
|
| 309 |
if stream:
|
| 310 |
kwargs.pop("decode_text")
|
|
|
|
| 273 |
else:
|
| 274 |
model_inputs["vision_hidden_states"] = vision_hidden_states
|
| 275 |
|
| 276 |
+
# print(model_inputs)
|
| 277 |
(
|
| 278 |
input_embeds,
|
| 279 |
vision_hidden_states_dummy,
|
|
|
|
| 304 |
print(stuff.shape, "SHAPE")
|
| 305 |
batch = torch.cat(padded_tensors, dim=0)
|
| 306 |
print(batch.shape)
|
| 307 |
+
print(batch)
|
| 308 |
# output_ids = self._decode(input_embeds, tokenizer, **kwargs)
|
| 309 |
if stream:
|
| 310 |
kwargs.pop("decode_text")
|