DylanZimmer commited on
Commit
444d328
·
1 Parent(s): bff1060

Print messages, history messing up is the guess

Browse files
Files changed (1) hide show
  1. app.py +14 -0
app.py CHANGED
@@ -39,6 +39,9 @@ def chat_fxn_caller(message, history, temperature=0.6, top_p=0.95, max_tokens=12
39
  """
40
  messages = []
41
 
 
 
 
42
  for turn in history:
43
  # Each turn is [user_msg, bot_msg] or [user_msg] depending on history length
44
  if isinstance(turn, list):
@@ -52,6 +55,9 @@ def chat_fxn_caller(message, history, temperature=0.6, top_p=0.95, max_tokens=12
52
  # Append current user message
53
  messages.append({"role": "user", "content": str(message)})
54
 
 
 
 
55
  prompt = chatbot.tokenizer.apply_chat_template(messages, tokenize=False)
56
 
57
  # Generate response
@@ -67,9 +73,17 @@ def chat_fxn_caller(message, history, temperature=0.6, top_p=0.95, max_tokens=12
67
  pad_token_id=chatbot.tokenizer.eos_token_id
68
  )
69
 
 
 
 
 
70
  # Extract only new text
71
  generated = response[0]["generated_text"][len(prompt):].strip()
72
 
 
 
 
 
73
  # Optional: truncate to avoid huge outputs on free GPU
74
  max_response_len = 512
75
  if len(generated) > max_response_len:
 
39
  """
40
  messages = []
41
 
42
+ print("history")
43
+ print(history)
44
+
45
  for turn in history:
46
  # Each turn is [user_msg, bot_msg] or [user_msg] depending on history length
47
  if isinstance(turn, list):
 
55
  # Append current user message
56
  messages.append({"role": "user", "content": str(message)})
57
 
58
+ print("messages")
59
+ print(messages)
60
+
61
  prompt = chatbot.tokenizer.apply_chat_template(messages, tokenize=False)
62
 
63
  # Generate response
 
73
  pad_token_id=chatbot.tokenizer.eos_token_id
74
  )
75
 
76
+ print("response:")
77
+ print(response)
78
+
79
+
80
  # Extract only new text
81
  generated = response[0]["generated_text"][len(prompt):].strip()
82
 
83
+ print("generated")
84
+ print(generated)
85
+
86
+
87
  # Optional: truncate to avoid huge outputs on free GPU
88
  max_response_len = 512
89
  if len(generated) > max_response_len: