removes truncation in tokenizer
Browse files
app.py
CHANGED
|
@@ -67,7 +67,7 @@ def run_llm(prompt: str, max_new_tokens: int = 64, temperature: float = 0.0, mod
|
|
| 67 |
tok, mdl = load_llm(model_id)
|
| 68 |
|
| 69 |
# Tokenize input
|
| 70 |
-
inputs = tok(prompt, return_tensors="pt"
|
| 71 |
inputs = {k: v.to(next(mdl.parameters()).device) for k, v in inputs.items()}
|
| 72 |
|
| 73 |
# Generate
|
|
|
|
| 67 |
tok, mdl = load_llm(model_id)
|
| 68 |
|
| 69 |
# Tokenize input
|
| 70 |
+
inputs = tok(prompt, return_tensors="pt")
|
| 71 |
inputs = {k: v.to(next(mdl.parameters()).device) for k, v in inputs.items()}
|
| 72 |
|
| 73 |
# Generate
|