Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -6,7 +6,7 @@ import os
|
|
| 6 |
auth_token = os.environ['HF_TOKEN']
|
| 7 |
|
| 8 |
# Load the tokenizer and models for the first pipeline
|
| 9 |
-
|
| 10 |
model_bin = AutoModelForTokenClassification.from_pretrained("AlGe/deberta-v3-large_token", token=auth_token)
|
| 11 |
tokenizer_ext.model_max_length = 512
|
| 12 |
pipe_bin = pipeline("ner", model=model_bin, tokenizer=tokenizer_bin)
|
|
|
|
| 6 |
auth_token = os.environ['HF_TOKEN']
|
| 7 |
|
| 8 |
# Load the tokenizer and models for the first pipeline
|
| 9 |
+
tokenizer_bin = AutoTokenizer.from_pretrained("AlGe/deberta-v3-large_token", token=auth_token)
|
| 10 |
model_bin = AutoModelForTokenClassification.from_pretrained("AlGe/deberta-v3-large_token", token=auth_token)
|
| 11 |
tokenizer_ext.model_max_length = 512
|
| 12 |
pipe_bin = pipeline("ner", model=model_bin, tokenizer=tokenizer_bin)
|