Spaces:
Sleeping
Sleeping
Upload app.py with huggingface_hub
Browse files
app.py
CHANGED
|
@@ -13,8 +13,8 @@ def get_pipeline():
|
|
| 13 |
if _PIPE is not None or _ERR is not None:
|
| 14 |
return _PIPE, _ERR
|
| 15 |
try:
|
| 16 |
-
tokenizer = AutoTokenizer.from_pretrained(MODEL_ID)
|
| 17 |
-
model = AutoModelForCausalLM.from_pretrained(MODEL_ID)
|
| 18 |
_PIPE = TextGenerationPipeline(model=model, tokenizer=tokenizer, device=-1)
|
| 19 |
except Exception as e:
|
| 20 |
_ERR = str(e)
|
|
|
|
| 13 |
if _PIPE is not None or _ERR is not None:
|
| 14 |
return _PIPE, _ERR
|
| 15 |
try:
|
| 16 |
+
tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, trust_remote_code=True)
|
| 17 |
+
model = AutoModelForCausalLM.from_pretrained(MODEL_ID, trust_remote_code=True, torch_dtype="auto")
|
| 18 |
_PIPE = TextGenerationPipeline(model=model, tokenizer=tokenizer, device=-1)
|
| 19 |
except Exception as e:
|
| 20 |
_ERR = str(e)
|