Spaces:
Paused
Paused
Update utils/llms.py
Browse files- utils/llms.py +42 -51
utils/llms.py
CHANGED
|
@@ -1,18 +1,27 @@
|
|
| 1 |
import os
|
| 2 |
import helpers.helper as helper
|
| 3 |
|
| 4 |
-
import google.generativeai as genai
|
| 5 |
from g4f.client import Client
|
| 6 |
from litellm import completion
|
| 7 |
import random
|
| 8 |
import json
|
|
|
|
|
|
|
| 9 |
|
| 10 |
-
|
|
|
|
| 11 |
|
| 12 |
gemini_api_keys=json.loads(os.environ.get("GEMINI_KEY_LIST"))
|
| 13 |
-
groq_api_keys=
|
|
|
|
|
|
|
|
|
|
| 14 |
DeepInfraChat.models = ["google/gemma-3-27b-it","deepseek-ai/DeepSeek-R1-Turbo","Qwen/QwQ-32B","deepseek-ai/DeepSeek-R1","deepseek-ai/DeepSeek-V3-0324","meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8","meta-llama/Llama-4-Scout-17B-16E-Instruct","microsoft/Phi-4-multimodal-instruct"]
|
|
|
|
| 15 |
deepinframodels=["meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8","microsoft/Phi-4-multimodal-instruct","google/gemma-3-27b-it","meta-llama/Llama-4-Scout-17B-16E-Instruct"]
|
|
|
|
|
|
|
|
|
|
| 16 |
REASONING_CORRESPONDANCE = {"DeepSeek-R1-Glider":Glider, "DeepSeekR1-LAMBDA":LambdaChat,"DeepSeekR1":DeepInfraChat,"deepseek-slow":TypeGPT}
|
| 17 |
os.environ["GEMINI_API_KEY"] =random.choice(gemini_api_keys)
|
| 18 |
|
|
@@ -26,20 +35,6 @@ clienty = OpenAI(
|
|
| 26 |
CHAT_CORRESPONDANCE = {"DeepSeek-V3":DeepInfraChat}
|
| 27 |
|
| 28 |
client = Client()
|
| 29 |
-
genai.configure(api_key="AIzaSyAQgAtQPpY0bQaCqCISGxeyF6tpDePx-Jg")
|
| 30 |
-
modell = genai.GenerativeModel('gemini-1.5-pro')
|
| 31 |
-
generation_config = {
|
| 32 |
-
"temperature": 1,
|
| 33 |
-
"top_p": 0.95,
|
| 34 |
-
"top_k": 40,
|
| 35 |
-
"max_output_tokens": 8192,
|
| 36 |
-
"response_mime_type": "text/plain",
|
| 37 |
-
}
|
| 38 |
-
|
| 39 |
-
model2flash = genai.GenerativeModel(
|
| 40 |
-
model_name="gemini-2.0-flash-thinking-exp",
|
| 41 |
-
generation_config=generation_config,
|
| 42 |
-
)
|
| 43 |
|
| 44 |
|
| 45 |
|
|
@@ -54,20 +49,12 @@ def clear():
|
|
| 54 |
helper.q.task_done()
|
| 55 |
|
| 56 |
def gpt4(messages,response_format,model="gpt-4"):
|
| 57 |
-
if response_format!=None:
|
| 58 |
-
output = clienty.chat.completions.create(
|
| 59 |
-
model="google/gemini-2.0-flash-001",
|
| 60 |
-
messages= messages,
|
| 61 |
-
response_format=response_format,
|
| 62 |
-
)
|
| 63 |
-
return str(output.choices[0].message.content)
|
| 64 |
-
else:
|
| 65 |
-
print(messages)
|
| 66 |
if len(messages) ==1:
|
| 67 |
messages[0]["role"]="user"
|
| 68 |
response = completion(
|
| 69 |
model="gemini/gemini-2.0-flash",
|
| 70 |
-
messages=messages
|
|
|
|
| 71 |
)
|
| 72 |
return str(response.choices[0].message.content)
|
| 73 |
|
|
@@ -87,7 +74,6 @@ def gpt4stream(messages,model,api_keys):
|
|
| 87 |
messages=messages,
|
| 88 |
stream=True
|
| 89 |
|
| 90 |
-
# Add any other necessary parameters
|
| 91 |
)
|
| 92 |
for part in response:
|
| 93 |
cunk=cunk+(part.choices[0].delta.content or "")
|
|
@@ -97,7 +83,6 @@ def gpt4stream(messages,model,api_keys):
|
|
| 97 |
clear()
|
| 98 |
break
|
| 99 |
except Exception as e:
|
| 100 |
-
|
| 101 |
pass
|
| 102 |
helper.q.put_nowait("RESULT: "+cunk)
|
| 103 |
|
|
@@ -128,7 +113,7 @@ def gpt4stream(messages,model,api_keys):
|
|
| 128 |
break
|
| 129 |
helper.q.put_nowait("RESULT: "+cunk)
|
| 130 |
|
| 131 |
-
elif model=="deepseek-r1-distill-llama-70b":
|
| 132 |
os.environ["GROQ_API_KEY"] =random.choice(groq_api_keys)
|
| 133 |
|
| 134 |
response = completion(model="groq/deepseek-r1-distill-llama-70b", messages=messages, stream=True)
|
|
@@ -142,7 +127,7 @@ def gpt4stream(messages,model,api_keys):
|
|
| 142 |
clear()
|
| 143 |
break
|
| 144 |
helper.q.put_nowait("RESULT: "+cunk)
|
| 145 |
-
elif model=="qwq-32b":
|
| 146 |
os.environ["GROQ_API_KEY"] =random.choice(groq_api_keys)
|
| 147 |
response = completion(model="groq/qwen-qwq-32b", messages=messages, stream=True)
|
| 148 |
|
|
@@ -169,34 +154,40 @@ def gpt4stream(messages,model,api_keys):
|
|
| 169 |
clear()
|
| 170 |
break
|
| 171 |
helper.q.put_nowait("RESULT: "+cunk)
|
|
|
|
|
|
|
| 172 |
|
| 173 |
-
|
| 174 |
-
for
|
| 175 |
-
|
| 176 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 177 |
|
| 178 |
-
|
|
|
|
| 179 |
|
| 180 |
-
|
| 181 |
-
|
| 182 |
-
|
| 183 |
-
|
| 184 |
-
|
| 185 |
-
|
| 186 |
-
|
| 187 |
-
|
| 188 |
-
break
|
| 189 |
-
except Exception as e:
|
| 190 |
-
helper.q.put_nowait(str(e))
|
| 191 |
-
pass
|
| 192 |
helper.q.put_nowait("RESULT: "+cunk)
|
| 193 |
|
| 194 |
-
elif
|
| 195 |
for key in gemini_api_keys:
|
| 196 |
try:
|
| 197 |
os.environ["GEMINI_API_KEY"] =key
|
| 198 |
|
| 199 |
-
response = completion(model="gemini/
|
| 200 |
|
| 201 |
cunk=""
|
| 202 |
for part in response:
|
|
@@ -251,7 +242,7 @@ def gpt4stream(messages,model,api_keys):
|
|
| 251 |
helper.q.put_nowait("RESULT: "+cunk)
|
| 252 |
|
| 253 |
|
| 254 |
-
elif model=="
|
| 255 |
helper.q.put_nowait("<think>")
|
| 256 |
cunk=""
|
| 257 |
providers=REASONING_QWQ
|
|
|
|
| 1 |
import os
|
| 2 |
import helpers.helper as helper
|
| 3 |
|
|
|
|
| 4 |
from g4f.client import Client
|
| 5 |
from litellm import completion
|
| 6 |
import random
|
| 7 |
import json
|
| 8 |
+
import os
|
| 9 |
+
# from dotenv import load_dotenv
|
| 10 |
|
| 11 |
+
# load_dotenv()
|
| 12 |
+
from g4f.Provider import DeepInfraChat,Glider,LambdaChat,TypeGPT
|
| 13 |
|
| 14 |
gemini_api_keys=json.loads(os.environ.get("GEMINI_KEY_LIST"))
|
| 15 |
+
groq_api_keys=json.loads(os.environ.get("GROQ_API_KEYS"))
|
| 16 |
+
chutes_key=os.environ.get("CHUTES_API_KEY")
|
| 17 |
+
github_key=os.environ.get("GITHUB_API_KEY")
|
| 18 |
+
|
| 19 |
DeepInfraChat.models = ["google/gemma-3-27b-it","deepseek-ai/DeepSeek-R1-Turbo","Qwen/QwQ-32B","deepseek-ai/DeepSeek-R1","deepseek-ai/DeepSeek-V3-0324","meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8","meta-llama/Llama-4-Scout-17B-16E-Instruct","microsoft/Phi-4-multimodal-instruct"]
|
| 20 |
+
|
| 21 |
deepinframodels=["meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8","microsoft/Phi-4-multimodal-instruct","google/gemma-3-27b-it","meta-llama/Llama-4-Scout-17B-16E-Instruct"]
|
| 22 |
+
chutes_models={"MAI-DS-R1-FP8":"microsoft/MAI-DS-R1-FP8","DeepSeek-V3-0324":"deepseek-ai/DeepSeek-V3-0324","deepseek-reasoner":"deepseek-ai/DeepSeek-R1","GLM-4-32B-0414":"THUDM/GLM-4-32B-0414","GLM-Z1-32B-0414":"THUDM/GLM-Z1-32B-0414"}
|
| 23 |
+
github_models={"gpt4.1":"gpt-4.1","gpt-4o":"gpt-4o","o4-mini":"o4-mini"}
|
| 24 |
+
|
| 25 |
REASONING_CORRESPONDANCE = {"DeepSeek-R1-Glider":Glider, "DeepSeekR1-LAMBDA":LambdaChat,"DeepSeekR1":DeepInfraChat,"deepseek-slow":TypeGPT}
|
| 26 |
os.environ["GEMINI_API_KEY"] =random.choice(gemini_api_keys)
|
| 27 |
|
|
|
|
| 35 |
CHAT_CORRESPONDANCE = {"DeepSeek-V3":DeepInfraChat}
|
| 36 |
|
| 37 |
client = Client()
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 38 |
|
| 39 |
|
| 40 |
|
|
|
|
| 49 |
helper.q.task_done()
|
| 50 |
|
| 51 |
def gpt4(messages,response_format,model="gpt-4"):
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 52 |
if len(messages) ==1:
|
| 53 |
messages[0]["role"]="user"
|
| 54 |
response = completion(
|
| 55 |
model="gemini/gemini-2.0-flash",
|
| 56 |
+
messages=messages,
|
| 57 |
+
response_format=response_format
|
| 58 |
)
|
| 59 |
return str(response.choices[0].message.content)
|
| 60 |
|
|
|
|
| 74 |
messages=messages,
|
| 75 |
stream=True
|
| 76 |
|
|
|
|
| 77 |
)
|
| 78 |
for part in response:
|
| 79 |
cunk=cunk+(part.choices[0].delta.content or "")
|
|
|
|
| 83 |
clear()
|
| 84 |
break
|
| 85 |
except Exception as e:
|
|
|
|
| 86 |
pass
|
| 87 |
helper.q.put_nowait("RESULT: "+cunk)
|
| 88 |
|
|
|
|
| 113 |
break
|
| 114 |
helper.q.put_nowait("RESULT: "+cunk)
|
| 115 |
|
| 116 |
+
elif model=="groq/deepseek-r1-distill-llama-70b":
|
| 117 |
os.environ["GROQ_API_KEY"] =random.choice(groq_api_keys)
|
| 118 |
|
| 119 |
response = completion(model="groq/deepseek-r1-distill-llama-70b", messages=messages, stream=True)
|
|
|
|
| 127 |
clear()
|
| 128 |
break
|
| 129 |
helper.q.put_nowait("RESULT: "+cunk)
|
| 130 |
+
elif model=="groq/qwq-32b":
|
| 131 |
os.environ["GROQ_API_KEY"] =random.choice(groq_api_keys)
|
| 132 |
response = completion(model="groq/qwen-qwq-32b", messages=messages, stream=True)
|
| 133 |
|
|
|
|
| 154 |
clear()
|
| 155 |
break
|
| 156 |
helper.q.put_nowait("RESULT: "+cunk)
|
| 157 |
+
elif model in chutes_models:
|
| 158 |
+
response = completion(model=f"openai/{chutes_models[model]}",api_key=chutes_key,base_url="https://llm.chutes.ai/v1", messages=messages, stream=True)
|
| 159 |
|
| 160 |
+
cunk=""
|
| 161 |
+
for part in response:
|
| 162 |
+
cunk=cunk+(part.choices[0].delta.content or "")
|
| 163 |
+
|
| 164 |
+
if "```json" not in cunk:
|
| 165 |
+
helper.q.put_nowait(part.choices[0].delta.content or "")
|
| 166 |
+
if helper.stopped:
|
| 167 |
+
clear()
|
| 168 |
+
break
|
| 169 |
+
helper.q.put_nowait("RESULT: "+cunk)
|
| 170 |
|
| 171 |
+
elif model in github_models:
|
| 172 |
+
response = completion(model=f"github/{github_models[model]}",api_key=github_key, messages=messages, stream=True)
|
| 173 |
|
| 174 |
+
cunk=""
|
| 175 |
+
for part in response:
|
| 176 |
+
cunk=cunk+(part.choices[0].delta.content or "")
|
| 177 |
+
|
| 178 |
+
if "```json" not in cunk:
|
| 179 |
+
helper.q.put_nowait(part.choices[0].delta.content or "")
|
| 180 |
+
if helper.stopped:
|
| 181 |
+
clear()
|
| 182 |
+
break
|
|
|
|
|
|
|
|
|
|
| 183 |
helper.q.put_nowait("RESULT: "+cunk)
|
| 184 |
|
| 185 |
+
elif "gemini" in model:
|
| 186 |
for key in gemini_api_keys:
|
| 187 |
try:
|
| 188 |
os.environ["GEMINI_API_KEY"] =key
|
| 189 |
|
| 190 |
+
response = completion(model=f"gemini/{model}", messages=messages, stream=True)
|
| 191 |
|
| 192 |
cunk=""
|
| 193 |
for part in response:
|
|
|
|
| 242 |
helper.q.put_nowait("RESULT: "+cunk)
|
| 243 |
|
| 244 |
|
| 245 |
+
elif model=="qwq-32b" :
|
| 246 |
helper.q.put_nowait("<think>")
|
| 247 |
cunk=""
|
| 248 |
providers=REASONING_QWQ
|