Spaces:
Sleeping
Sleeping
Update tasks/text.py
Browse files- tasks/text.py +6 -1
tasks/text.py
CHANGED
|
@@ -84,7 +84,12 @@ async def evaluate_text(request: TextEvaluationRequest):
|
|
| 84 |
model = PeftModel.from_pretrained(base_model, path_adapter)
|
| 85 |
model.eval()
|
| 86 |
tokenizer = AutoTokenizer.from_pretrained(path_model)
|
| 87 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 88 |
def preprocess_function(df):
|
| 89 |
return tokenizer(df["quote"], truncation=True)
|
| 90 |
tokenized_test = test_dataset.map(preprocess_function, batched=True)
|
|
|
|
| 84 |
model = PeftModel.from_pretrained(base_model, path_adapter)
|
| 85 |
model.eval()
|
| 86 |
tokenizer = AutoTokenizer.from_pretrained(path_model)
|
| 87 |
+
tokenizer.pad_token = tokenizer.eos_token # Or any other token depending on your model
|
| 88 |
+
tokenizer.pad_token_id = tokenizer.eos_token_id
|
| 89 |
+
|
| 90 |
+
model.config.pad_token_id = tokenizer.pad_token_id
|
| 91 |
+
model.config.use_cache = False
|
| 92 |
+
model.config.pretraining_tp = 1
|
| 93 |
def preprocess_function(df):
|
| 94 |
return tokenizer(df["quote"], truncation=True)
|
| 95 |
tokenized_test = test_dataset.map(preprocess_function, batched=True)
|