airsltd commited on
Commit
c124a7b
·
verified ·
1 Parent(s): 05e9938

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -12
app.py CHANGED
@@ -24,8 +24,8 @@ def check_and_download_model():
24
 
25
  # Use TinyLlama - a fully public model
26
  # model_name = "TinyLlama/TinyLlama-1.1B-Chat-v1.0"
27
- # model_name = "unsloth/functiongemma-270m-it"
28
- model_name = "Qwen/Qwen3-0.6B"
29
  cache_dir = "./my_model_cache"
30
 
31
  # Check if model already exists in cache
@@ -142,7 +142,7 @@ def openai_chat_completions(request: dict):
142
 
143
  messages = request.get("messages", [])
144
  model = request.get("model", model_name)
145
- max_tokens = request.get("max_tokens", 100)
146
  temperature = request.get("temperature", 0.7)
147
 
148
  print('\n\n request')
@@ -162,17 +162,8 @@ def openai_chat_completions(request: dict):
162
  max_new_tokens=max_tokens,
163
  # temperature=temperature
164
  )
165
- print('\n\n asdfasdfasdfasdf')
166
- print('result')
167
- print(result)
168
- print('data over \n\n')
169
-
170
 
171
  result = convert_json_format(result)
172
- print('\n\n 33333333333333333333333')
173
- print('result')
174
- print(result)
175
- print('data over 3333 \n\n')
176
 
177
 
178
  completion_id = f"chatcmpl-{int(time.time())}"
 
24
 
25
  # Use TinyLlama - a fully public model
26
  # model_name = "TinyLlama/TinyLlama-1.1B-Chat-v1.0"
27
+ model_name = "unsloth/functiongemma-270m-it"
28
+ # model_name = "Qwen/Qwen3-0.6B"
29
  cache_dir = "./my_model_cache"
30
 
31
  # Check if model already exists in cache
 
142
 
143
  messages = request.get("messages", [])
144
  model = request.get("model", model_name)
145
+ max_tokens = request.get("max_tokens", 1000)
146
  temperature = request.get("temperature", 0.7)
147
 
148
  print('\n\n request')
 
162
  max_new_tokens=max_tokens,
163
  # temperature=temperature
164
  )
 
 
 
 
 
165
 
166
  result = convert_json_format(result)
 
 
 
 
167
 
168
 
169
  completion_id = f"chatcmpl-{int(time.time())}"