Update README.md
Browse files
README.md
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
---
|
| 2 |
base_model:
|
| 3 |
-
- unsloth/Llama-3.2-1B
|
| 4 |
library_name: peft
|
| 5 |
license: apache-2.0
|
| 6 |
language:
|
|
@@ -35,8 +35,8 @@ device = "cuda" if torch.cuda.is_available() else "cpu"
|
|
| 35 |
|
| 36 |
# Load tokenizer and models
|
| 37 |
print("Loading models...")
|
| 38 |
-
tokenizer = AutoTokenizer.from_pretrained("unsloth/llama-3.2-1b-instruct
|
| 39 |
-
base_model = AutoModelForCausalLM.from_pretrained("unsloth/llama-3.2-1b-instruct
|
| 40 |
model = PeftModel.from_pretrained(base_model, "HackWeasel/llama-3.2-1b-QLORA-IMDB").to(device)
|
| 41 |
model.eval()
|
| 42 |
print("Models loaded!")
|
|
|
|
| 1 |
---
|
| 2 |
base_model:
|
| 3 |
+
- unsloth/Llama-3.2-1B-Instruct
|
| 4 |
library_name: peft
|
| 5 |
license: apache-2.0
|
| 6 |
language:
|
|
|
|
| 35 |
|
| 36 |
# Load tokenizer and models
|
| 37 |
print("Loading models...")
|
| 38 |
+
tokenizer = AutoTokenizer.from_pretrained("unsloth/llama-3.2-1b-instruct")
|
| 39 |
+
base_model = AutoModelForCausalLM.from_pretrained("unsloth/llama-3.2-1b-instruct").to(device)
|
| 40 |
model = PeftModel.from_pretrained(base_model, "HackWeasel/llama-3.2-1b-QLORA-IMDB").to(device)
|
| 41 |
model.eval()
|
| 42 |
print("Models loaded!")
|