Update README.md
Browse files
README.md
CHANGED
|
@@ -35,8 +35,8 @@ device = "cuda" if torch.cuda.is_available() else "cpu"
|
|
| 35 |
|
| 36 |
# Load tokenizer and models
|
| 37 |
print("Loading models...")
|
| 38 |
-
tokenizer = AutoTokenizer.from_pretrained("unsloth/llama-3.2-1b-instruct")
|
| 39 |
-
base_model = AutoModelForCausalLM.from_pretrained("unsloth/llama-3.2-1b-instruct").to(device)
|
| 40 |
model = PeftModel.from_pretrained(base_model, "HackWeasel/llama-3.2-1b-QLORA-IMDB").to(device)
|
| 41 |
model.eval()
|
| 42 |
print("Models loaded!")
|
|
|
|
| 35 |
|
| 36 |
# Load tokenizer and models
|
| 37 |
print("Loading models...")
|
| 38 |
+
tokenizer = AutoTokenizer.from_pretrained("unsloth/llama-3.2-1b-instruct-bnb-4bit")
|
| 39 |
+
base_model = AutoModelForCausalLM.from_pretrained("unsloth/llama-3.2-1b-instruct-bnb-4bit").to(device)
|
| 40 |
model = PeftModel.from_pretrained(base_model, "HackWeasel/llama-3.2-1b-QLORA-IMDB").to(device)
|
| 41 |
model.eval()
|
| 42 |
print("Models loaded!")
|