Update README.md
Browse files
README.md
CHANGED
|
@@ -180,7 +180,7 @@ import torch
|
|
| 180 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 181 |
|
| 182 |
# Load model and tokenizer
|
| 183 |
-
model_name = "
|
| 184 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 185 |
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16, device_map="auto")
|
| 186 |
|
|
@@ -221,7 +221,7 @@ import torch
|
|
| 221 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 222 |
|
| 223 |
class CharacterChat:
|
| 224 |
-
def __init__(self, model_path="
|
| 225 |
print(f"Loading model from {model_path}...")
|
| 226 |
self.tokenizer = AutoTokenizer.from_pretrained(model_path)
|
| 227 |
self.model = AutoModelForCausalLM.from_pretrained(
|
|
|
|
| 180 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 181 |
|
| 182 |
# Load model and tokenizer
|
| 183 |
+
model_name = "luvgpt/phi3-uncensored-chat"
|
| 184 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 185 |
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16, device_map="auto")
|
| 186 |
|
|
|
|
| 221 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 222 |
|
| 223 |
class CharacterChat:
|
| 224 |
+
def __init__(self, model_path="luvgpt/phi3-uncensored-chat", persona=None):
|
| 225 |
print(f"Loading model from {model_path}...")
|
| 226 |
self.tokenizer = AutoTokenizer.from_pretrained(model_path)
|
| 227 |
self.model = AutoModelForCausalLM.from_pretrained(
|