Fix code typo for Simple and Video Inference examples
#9
by
kubistmi
- opened
README.md
CHANGED
|
@@ -106,7 +106,7 @@ inputs = processor.apply_chat_template(
|
|
| 106 |
tokenize=True,
|
| 107 |
return_dict=True,
|
| 108 |
return_tensors="pt",
|
| 109 |
-
).to(model.device)
|
| 110 |
|
| 111 |
generated_ids = model.generate(**inputs, do_sample=False, max_new_tokens=64)
|
| 112 |
generated_texts = processor.batch_decode(
|
|
@@ -137,7 +137,7 @@ inputs = processor.apply_chat_template(
|
|
| 137 |
tokenize=True,
|
| 138 |
return_dict=True,
|
| 139 |
return_tensors="pt",
|
| 140 |
-
).to(model.device)
|
| 141 |
|
| 142 |
generated_ids = model.generate(**inputs, do_sample=False, max_new_tokens=64)
|
| 143 |
generated_texts = processor.batch_decode(
|
|
|
|
| 106 |
tokenize=True,
|
| 107 |
return_dict=True,
|
| 108 |
return_tensors="pt",
|
| 109 |
+
).to(model.device, dtype=torch.bfloat16)
|
| 110 |
|
| 111 |
generated_ids = model.generate(**inputs, do_sample=False, max_new_tokens=64)
|
| 112 |
generated_texts = processor.batch_decode(
|
|
|
|
| 137 |
tokenize=True,
|
| 138 |
return_dict=True,
|
| 139 |
return_tensors="pt",
|
| 140 |
+
).to(model.device, dtype=torch.bfloat16)
|
| 141 |
|
| 142 |
generated_ids = model.generate(**inputs, do_sample=False, max_new_tokens=64)
|
| 143 |
generated_texts = processor.batch_decode(
|