Spaces:
Sleeping
Sleeping
Change model to 8B
Browse files
app.py
CHANGED
|
@@ -14,7 +14,7 @@ hf_token = os.getenv("HF_TOKEN")
|
|
| 14 |
|
| 15 |
|
| 16 |
# --- Load tokenizer ---
|
| 17 |
-
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.
|
| 18 |
vocab_size = len(tokenizer)
|
| 19 |
eos_token_id = tokenizer.eos_token_id
|
| 20 |
mask_token_id = tokenizer.encode('MASK', add_special_tokens=False)[0]
|
|
@@ -37,9 +37,9 @@ assistant_marker_ids = tokenizer.encode("Assistant:", add_special_tokens=False)
|
|
| 37 |
def load_model():
|
| 38 |
ckpt_path = hf_hub_download(
|
| 39 |
repo_id="ruurd/tini_model",
|
| 40 |
-
filename="diffusion-model.pth",
|
| 41 |
token=os.getenv("HF_TOKEN"),
|
| 42 |
-
revision="1ffb916dd34f442f87cf06dda74b96f86eaf1d15",
|
| 43 |
)
|
| 44 |
|
| 45 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
|
|
|
| 14 |
|
| 15 |
|
| 16 |
# --- Load tokenizer ---
|
| 17 |
+
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.1-8B", use_fast=True, token=hf_token)
|
| 18 |
vocab_size = len(tokenizer)
|
| 19 |
eos_token_id = tokenizer.eos_token_id
|
| 20 |
mask_token_id = tokenizer.encode('MASK', add_special_tokens=False)[0]
|
|
|
|
| 37 |
def load_model():
|
| 38 |
ckpt_path = hf_hub_download(
|
| 39 |
repo_id="ruurd/tini_model",
|
| 40 |
+
filename="diffusion-model-8B.pth",
|
| 41 |
token=os.getenv("HF_TOKEN"),
|
| 42 |
+
# revision="1ffb916dd34f442f87cf06dda74b96f86eaf1d15",
|
| 43 |
)
|
| 44 |
|
| 45 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|