Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -21,7 +21,7 @@ MAX_INPUT_TOKEN_LENGTH = 32000
|
|
| 21 |
|
| 22 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
| 23 |
|
| 24 |
-
tokenizer = AutoTokenizer.from_pretrained("EvaByte/EvaByte", trust_remote_code=True)
|
| 25 |
model = AutoModelForCausalLM.from_pretrained(
|
| 26 |
"EvaByte/EvaByte-SFT",
|
| 27 |
torch_dtype=torch.bfloat16,
|
|
|
|
| 21 |
|
| 22 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
| 23 |
|
| 24 |
+
tokenizer = AutoTokenizer.from_pretrained("EvaByte/EvaByte-SFT", trust_remote_code=True)
|
| 25 |
model = AutoModelForCausalLM.from_pretrained(
|
| 26 |
"EvaByte/EvaByte-SFT",
|
| 27 |
torch_dtype=torch.bfloat16,
|