Spaces:
Sleeping
Sleeping
Change model_base_id and lora_path
Browse files
app.py
CHANGED
|
@@ -20,8 +20,8 @@ from transformers import AutoTokenizer, AutoModelForCausalLM
|
|
| 20 |
from peft import PeftModel
|
| 21 |
|
| 22 |
# === [1] Model and Tokenizer Loading ===
|
| 23 |
-
base_model_id = "
|
| 24 |
-
lora_path = "./tat-llm" # Path to your fine-tuned LoRA folder
|
| 25 |
|
| 26 |
# Load base model and LoRA adapter
|
| 27 |
base_model = AutoModelForCausalLM.from_pretrained(base_model_id, torch_dtype=torch.float16)
|
|
|
|
| 20 |
from peft import PeftModel
|
| 21 |
|
| 22 |
# === [1] Model and Tokenizer Loading ===
|
| 23 |
+
base_model_id = "NousResearch/Nous-Hermes-2-Mistral-7B-DPO"
|
| 24 |
+
lora_path = "./tat-llm-final-e3" # Path to your fine-tuned LoRA folder
|
| 25 |
|
| 26 |
# Load base model and LoRA adapter
|
| 27 |
base_model = AutoModelForCausalLM.from_pretrained(base_model_id, torch_dtype=torch.float16)
|