Ilke Ileri commited on
Commit
e683a4c
·
1 Parent(s): 25839d0

Improve response formatting and generation parameters

Browse files
Files changed (1) hide show
  1. app.py +10 -4
app.py CHANGED
@@ -66,15 +66,21 @@ def chat_completions():
66
  inputs = tokenizer(formatted_prompt, return_tensors="pt")
67
  outputs = model.generate(
68
  inputs["input_ids"],
69
- max_new_tokens=200,
70
  temperature=0.7,
71
  do_sample=True,
72
  top_p=0.9,
73
- num_return_sequences=1
 
 
 
74
  )
75
 
76
- full_response = tokenizer.decode(outputs[0], skip_special_tokens=True)
77
- response_text = full_response.replace(formatted_prompt, "").strip()
 
 
 
78
 
79
  # Vapi formatı
80
  vapi_response = {
 
66
  inputs = tokenizer(formatted_prompt, return_tensors="pt")
67
  outputs = model.generate(
68
  inputs["input_ids"],
69
+ max_new_tokens=256,
70
  temperature=0.7,
71
  do_sample=True,
72
  top_p=0.9,
73
+ num_return_sequences=1,
74
+ repetition_penalty=1.15,
75
+ pad_token_id=tokenizer.pad_token_id,
76
+ eos_token_id=tokenizer.eos_token_id
77
  )
78
 
79
+ full_response = tokenizer.decode(outputs[0], skip_special_tokens=False)
80
+
81
+ # Gemma formatındaki tag'leri temizle
82
+ response_text = full_response.split("<start_of_turn>model\n")[-1]
83
+ response_text = response_text.replace("<end_of_turn>", "").strip()
84
 
85
  # Vapi formatı
86
  vapi_response = {