Spaces:
Sleeping
Sleeping
Update src/plan.py
Browse files- src/plan.py +2 -7
src/plan.py
CHANGED
|
@@ -4,15 +4,10 @@ import json
|
|
| 4 |
import re
|
| 5 |
import os
|
| 6 |
|
| 7 |
-
model_id = "google/gemma-3n-E4B-it"
|
| 8 |
-
|
| 9 |
-
# Set Hugging Face cache directory
|
| 10 |
-
HF_CACHE_DIR = "./hf_cache"
|
| 11 |
-
os.environ["HF_HOME"] = HF_CACHE_DIR
|
| 12 |
-
os.environ["TRANSFORMERS_CACHE"] = HF_CACHE_DIR
|
| 13 |
-
os.makedirs(HF_CACHE_DIR, exist_ok=True)
|
| 14 |
|
| 15 |
hf_token = os.environ.get("HUGGINGFACE_TOKEN")
|
|
|
|
|
|
|
| 16 |
tokenizer = AutoTokenizer.from_pretrained(model_id, token=hf_token, use_auth_token=True)
|
| 17 |
model = AutoModelForCausalLM.from_pretrained(model_id, token=hf_token)
|
| 18 |
|
|
|
|
| 4 |
import re
|
| 5 |
import os
|
| 6 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 7 |
|
| 8 |
hf_token = os.environ.get("HUGGINGFACE_TOKEN")
|
| 9 |
+
model_id = "google/gemma-3n-E4B-it"
|
| 10 |
+
|
| 11 |
tokenizer = AutoTokenizer.from_pretrained(model_id, token=hf_token, use_auth_token=True)
|
| 12 |
model = AutoModelForCausalLM.from_pretrained(model_id, token=hf_token)
|
| 13 |
|