Update app.py
Browse files
app.py
CHANGED
|
@@ -22,8 +22,10 @@ model = AutoModelForCausalLM.from_pretrained("justinj92/phi3-orpo")
|
|
| 22 |
|
| 23 |
terminators = [
|
| 24 |
tok.eos_token_id,
|
| 25 |
-
|
| 26 |
-
|
|
|
|
|
|
|
| 27 |
]
|
| 28 |
|
| 29 |
if torch.cuda.is_available():
|
|
@@ -39,7 +41,7 @@ model = model.to(device)
|
|
| 39 |
|
| 40 |
@spaces.GPU(duration=60)
|
| 41 |
def chat(message, history, temperature, do_sample, max_tokens):
|
| 42 |
-
chat = []
|
| 43 |
for item in history:
|
| 44 |
chat.append({"role": "user", "content": item[0]})
|
| 45 |
if item[1] is not None:
|
|
|
|
| 22 |
|
| 23 |
terminators = [
|
| 24 |
tok.eos_token_id,
|
| 25 |
+
32007,
|
| 26 |
+
32011,
|
| 27 |
+
32001,
|
| 28 |
+
32000
|
| 29 |
]
|
| 30 |
|
| 31 |
if torch.cuda.is_available():
|
|
|
|
| 41 |
|
| 42 |
@spaces.GPU(duration=60)
|
| 43 |
def chat(message, history, temperature, do_sample, max_tokens):
|
| 44 |
+
chat = [{"role": "system", "content": "You are ORPO Tuned Phi Beast. Answer all questions in the most helpful way. No yapping."}]
|
| 45 |
for item in history:
|
| 46 |
chat.append({"role": "user", "content": item[0]})
|
| 47 |
if item[1] is not None:
|