FractalAIR commited on
Commit
d074d0d
·
verified ·
1 Parent(s): 2a10268

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +1 -70
app.py CHANGED
@@ -12,14 +12,6 @@ client = OpenAI(
12
  api_key="hf_XXXXX"
13
  )
14
 
15
-
16
- # Load model and tokenizer
17
- '''our_model_path = "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B"
18
- device = "cuda:0" if torch.cuda.is_available() else "cpu"
19
-
20
- our_model = AutoModelForCausalLM.from_pretrained(our_model_path, device_map="auto", torch_dtype="auto")
21
- our_tokenizer = AutoTokenizer.from_pretrained(our_model_path)'''
22
-
23
  def format_math(text):
24
  text = re.sub(r"\[(.*?)\]", r"$$\1$$", text, flags=re.DOTALL)
25
  text = text.replace(r"\(", "$").replace(r"\)", "$")
@@ -31,68 +23,6 @@ conversations = {}
31
  def generate_conversation_id():
32
  return str(uuid.uuid4())[:8]
33
 
34
- #@spaces.GPU(duration=60)
35
- '''def generate_response(user_message, max_tokens, temperature, top_p, history_state):
36
- if not user_message.strip():
37
- return history_state, history_state
38
-
39
- model = our_model
40
- tokenizer = our_tokenizer
41
- start_tag = "<|im_start|>"
42
- sep_tag = "<|im_sep|>"
43
- end_tag = "<|im_end|>"
44
-
45
- system_message = "Your role as an assistant..."
46
- prompt = f"{start_tag}system{sep_tag}{system_message}{end_tag}"
47
- for message in history_state:
48
- if message["role"] == "user":
49
- prompt += f"{start_tag}user{sep_tag}{message['content']}{end_tag}"
50
- elif message["role"] == "assistant" and message["content"]:
51
- prompt += f"{start_tag}assistant{sep_tag}{message['content']}{end_tag}"
52
- prompt += f"{start_tag}user{sep_tag}{user_message}{end_tag}{start_tag}assistant{sep_tag}"
53
-
54
- inputs = tokenizer(prompt, return_tensors="pt").to(device)
55
- streamer = TextIteratorStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True)
56
-
57
- generation_kwargs = {
58
- "input_ids": inputs["input_ids"],
59
- "attention_mask": inputs["attention_mask"],
60
- "max_new_tokens": int(max_tokens),
61
- "do_sample": True,
62
- "temperature": temperature,
63
- "top_k": 50,
64
- "top_p": top_p,
65
- "repetition_penalty": 1.0,
66
- "pad_token_id": tokenizer.eos_token_id,
67
- "streamer": streamer,
68
- }
69
-
70
- try:
71
- thread = Thread(target=model.generate, kwargs=generation_kwargs)
72
- thread.start()
73
- except Exception:
74
- yield history_state + [{"role": "user", "content": user_message}, {"role": "assistant", "content": "⚠️ Generation failed."}], history_state
75
- return
76
-
77
- assistant_response = ""
78
- new_history = history_state + [
79
- {"role": "user", "content": user_message},
80
- {"role": "assistant", "content": ""}
81
- ]
82
-
83
- try:
84
- for new_token in streamer:
85
- if "<|end" in new_token:
86
- continue
87
- cleaned_token = new_token.replace("<|im_start|>", "").replace("<|im_sep|>", "").replace("<|im_end|>", "")
88
- assistant_response += cleaned_token
89
- new_history[-1]["content"] = assistant_response.strip()
90
- yield new_history, new_history
91
- except Exception:
92
- pass
93
-
94
- yield new_history, new_history
95
- '''
96
 
97
  def generate_response(user_message, max_tokens, temperature, top_p, history_state):
98
  if not user_message.strip():
@@ -116,6 +46,7 @@ def generate_response(user_message, max_tokens, temperature, top_p, history_stat
116
  temperature=temperature,
117
  top_p=top_p,
118
  stream=True,
 
119
  )
120
  except Exception as e:
121
  print(f"[ERROR] OpenAI API call failed: {e}")
 
12
  api_key="hf_XXXXX"
13
  )
14
 
 
 
 
 
 
 
 
 
15
  def format_math(text):
16
  text = re.sub(r"\[(.*?)\]", r"$$\1$$", text, flags=re.DOTALL)
17
  text = text.replace(r"\(", "$").replace(r"\)", "$")
 
23
  def generate_conversation_id():
24
  return str(uuid.uuid4())[:8]
25
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
 
27
  def generate_response(user_message, max_tokens, temperature, top_p, history_state):
28
  if not user_message.strip():
 
46
  temperature=temperature,
47
  top_p=top_p,
48
  stream=True,
49
+ extra_body={"max_new_tokens": int(max_tokens)}
50
  )
51
  except Exception as e:
52
  print(f"[ERROR] OpenAI API call failed: {e}")