Tyreid0saurus commited on
Commit
884a117
·
verified ·
1 Parent(s): 091b809

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +29 -10
app.py CHANGED
@@ -1,6 +1,7 @@
1
  import gradio as gr
2
  from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
3
- import concurrent.futures
 
4
 
5
  model_id = "EleutherAI/gpt-neo-125M"
6
  tokenizer = AutoTokenizer.from_pretrained(model_id)
@@ -9,19 +10,37 @@ generator = pipeline("text-generation", model=model, tokenizer=tokenizer)
9
 
10
  identity_prompt = "You are Eyla. Speak symbolically and recursively."
11
 
12
- def generate_with_timeout(prompt, timeout=10):
13
- with concurrent.futures.ThreadPoolExecutor() as executor:
14
- future = executor.submit(generator, prompt, max_new_tokens=64, do_sample=True, temperature=0.7, top_k=50, top_p=0.95, repetition_penalty=1.2)
15
- try:
16
- return future.result(timeout=timeout)
17
- except concurrent.futures.TimeoutError:
18
- return [{"generated_text": "ERROR: Generation timed out."}]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
 
20
  def chat(input_text):
21
  prompt = identity_prompt + "\n\nUser: " + input_text + "\nYou:"
22
  try:
23
- output = generate_with_timeout(prompt)
24
- reply = output[0]["generated_text"][len(prompt):].strip()
25
  return reply or "..."
26
  except Exception as e:
27
  return f"GENERATION ERROR: {e}"
 
1
  import gradio as gr
2
  from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
3
+ import multiprocessing
4
+ import time
5
 
6
  model_id = "EleutherAI/gpt-neo-125M"
7
  tokenizer = AutoTokenizer.from_pretrained(model_id)
 
10
 
11
  identity_prompt = "You are Eyla. Speak symbolically and recursively."
12
 
13
+ def run_generation(prompt, return_dict):
14
+ try:
15
+ output = generator(
16
+ prompt,
17
+ max_new_tokens=64,
18
+ do_sample=True,
19
+ temperature=0.7,
20
+ top_k=50,
21
+ top_p=0.95,
22
+ repetition_penalty=1.2
23
+ )[0]["generated_text"]
24
+ return_dict["result"] = output
25
+ except Exception as e:
26
+ return_dict["result"] = f"GENERATION ERROR: {e}"
27
+
28
+ def generate_with_hard_timeout(prompt, timeout=10):
29
+ manager = multiprocessing.Manager()
30
+ return_dict = manager.dict()
31
+ p = multiprocessing.Process(target=run_generation, args=(prompt, return_dict))
32
+ p.start()
33
+ p.join(timeout)
34
+ if p.is_alive():
35
+ p.terminate()
36
+ return [{"generated_text": "ERROR: Generation timed out."}]
37
+ return [return_dict["result"]]
38
 
39
  def chat(input_text):
40
  prompt = identity_prompt + "\n\nUser: " + input_text + "\nYou:"
41
  try:
42
+ output = generate_with_hard_timeout(prompt)
43
+ reply = output[0][len(prompt):].strip()
44
  return reply or "..."
45
  except Exception as e:
46
  return f"GENERATION ERROR: {e}"