jowilke77 commited on
Commit
2eb7fef
·
verified ·
1 Parent(s): a4b7286

Create inference.py

Browse files
Files changed (1) hide show
  1. inference.py +32 -0
inference.py ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from transformers import AutoTokenizer, AutoModelForCausalLM
2
+ import torch
3
+
4
+ MODEL_NAME = "distilgpt2"
5
+
6
+ tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
7
+ model = AutoModelForCausalLM.from_pretrained(MODEL_NAME)
8
+
9
+ with open("prompt.txt", "r") as f:
10
+ SYSTEM_PROMPT = f.read().strip()
11
+
12
+ def generate(user_input):
13
+ prompt = f"{SYSTEM_PROMPT}\n\nUser: {user_input}\nBrad AI:"
14
+ inputs = tokenizer(prompt, return_tensors="pt")
15
+
16
+ with torch.no_grad():
17
+ outputs = model.generate(
18
+ **inputs,
19
+ max_new_tokens=150,
20
+ temperature=0.7,
21
+ top_p=0.9,
22
+ do_sample=True
23
+ )
24
+
25
+ return tokenizer.decode(outputs[0], skip_special_tokens=True)
26
+
27
+ if __name__ == "__main__":
28
+ while True:
29
+ user = input("You: ")
30
+ if user.lower() in ["exit", "quit"]:
31
+ break
32
+ print(generate(user))