AxionLab-official commited on
Commit
e3d8050
·
verified ·
1 Parent(s): 6466616

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +16 -16
app.py CHANGED
@@ -52,24 +52,24 @@ def chat(user_input, system_prompt, temperature, top_p, max_tokens):
52
  ).to(model.device)
53
 
54
  with torch.no_grad():
55
- output = model.generate(
56
- **inputs,
57
- max_new_tokens=int(max_tokens),
58
- do_sample=True,
59
- temperature=float(temperature),
60
- top_p=float(top_p),
61
- repetition_penalty=1.1,
62
- eos_token_id=tokenizer.eos_token_id,
63
- pad_token_id=tokenizer.eos_token_id
64
- )
65
-
66
- generated_tokens = output[0][inputs["input_ids"].shape[-1]:]
67
-
68
- decoded = tokenizer.decode(
69
- generated_tokens,
70
- skip_special_tokens=True
71
  )
72
 
 
 
 
 
 
 
 
73
  return decoded.strip()
74
 
75
  # =========================
 
52
  ).to(model.device)
53
 
54
  with torch.no_grad():
55
+ output = model.generate(
56
+ **inputs,
57
+ max_new_tokens=int(max_tokens),
58
+ do_sample=True,
59
+ temperature=float(temperature),
60
+ top_p=float(top_p),
61
+ repetition_penalty=1.1,
62
+ eos_token_id=tokenizer.eos_token_id,
63
+ pad_token_id=tokenizer.eos_token_id
 
 
 
 
 
 
 
64
  )
65
 
66
+ generated_tokens = output[0][inputs["input_ids"].shape[-1]:]
67
+
68
+ decoded = tokenizer.decode(
69
+ generated_tokens,
70
+ skip_special_tokens=True
71
+ )
72
+
73
  return decoded.strip()
74
 
75
  # =========================