Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -1,7 +1,11 @@
|
|
| 1 |
import gradio as gr
|
| 2 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
|
|
| 3 |
|
| 4 |
-
|
|
|
|
|
|
|
|
|
|
| 5 |
tokenizer = AutoTokenizer.from_pretrained("ewernn/perfect-refusal-model")
|
| 6 |
|
| 7 |
def chat(message):
|
|
|
|
| 1 |
import gradio as gr
|
| 2 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 3 |
+
from peft import PeftModel
|
| 4 |
|
| 5 |
+
# Load base model
|
| 6 |
+
base_model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen2.5-0.5B-Instruct")
|
| 7 |
+
# Load LoRA adapters on top
|
| 8 |
+
model = PeftModel.from_pretrained(base_model, "ewernn/perfect-refusal-model")
|
| 9 |
tokenizer = AutoTokenizer.from_pretrained("ewernn/perfect-refusal-model")
|
| 10 |
|
| 11 |
def chat(message):
|