Update app.py
Browse files
app.py
CHANGED
|
@@ -2,7 +2,7 @@ import os
|
|
| 2 |
import gradio as gr
|
| 3 |
import torch
|
| 4 |
import json
|
| 5 |
-
from transformers import
|
| 6 |
from peft import PeftModel
|
| 7 |
|
| 8 |
# Set Hugging Face Token for Authentication
|
|
@@ -26,7 +26,7 @@ LLAMA_GUARD_NAME = "meta-llama/Llama-Guard-3-1B-INT4" # Ensure this is correct
|
|
| 26 |
def load_llama_model(base_model=BASE_MODEL, adapter=None):
|
| 27 |
print(f"🔄 Loading Base Model: {base_model}")
|
| 28 |
|
| 29 |
-
tokenizer =
|
| 30 |
model = AutoModelForCausalLM.from_pretrained(
|
| 31 |
base_model,
|
| 32 |
token=HUGGINGFACE_TOKEN,
|
|
|
|
| 2 |
import gradio as gr
|
| 3 |
import torch
|
| 4 |
import json
|
| 5 |
+
from transformers import LlamaTokenizer, AutoModelForCausalLM
|
| 6 |
from peft import PeftModel
|
| 7 |
|
| 8 |
# Set Hugging Face Token for Authentication
|
|
|
|
| 26 |
def load_llama_model(base_model=BASE_MODEL, adapter=None):
|
| 27 |
print(f"🔄 Loading Base Model: {base_model}")
|
| 28 |
|
| 29 |
+
tokenizer = LlamaTokenizer.from_pretrained(base_model, token=HUGGINGFACE_TOKEN)
|
| 30 |
model = AutoModelForCausalLM.from_pretrained(
|
| 31 |
base_model,
|
| 32 |
token=HUGGINGFACE_TOKEN,
|