Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -192,7 +192,6 @@ def inference_batch_transformers(
|
|
| 192 |
).to(device)
|
| 193 |
|
| 194 |
speech_end_id = tok.convert_tokens_to_ids("<|SPEECH_GENERATION_END|>")
|
| 195 |
-
pad_id = tok.pad_token_id if tok.pad_token_id is not None else (tok.eos_token_id or speech_end_id)
|
| 196 |
|
| 197 |
outputs = lm.generate(
|
| 198 |
input_ids=input_ids_1,
|
|
@@ -203,7 +202,6 @@ def inference_batch_transformers(
|
|
| 203 |
repetition_penalty=1.05,
|
| 204 |
max_new_tokens=max_new_tokens,
|
| 205 |
eos_token_id=speech_end_id,
|
| 206 |
-
pad_token_id=pad_id,
|
| 207 |
num_return_sequences=batch_size,
|
| 208 |
use_cache=True,
|
| 209 |
)
|
|
|
|
| 192 |
).to(device)
|
| 193 |
|
| 194 |
speech_end_id = tok.convert_tokens_to_ids("<|SPEECH_GENERATION_END|>")
|
|
|
|
| 195 |
|
| 196 |
outputs = lm.generate(
|
| 197 |
input_ids=input_ids_1,
|
|
|
|
| 202 |
repetition_penalty=1.05,
|
| 203 |
max_new_tokens=max_new_tokens,
|
| 204 |
eos_token_id=speech_end_id,
|
|
|
|
| 205 |
num_return_sequences=batch_size,
|
| 206 |
use_cache=True,
|
| 207 |
)
|