binary1ne commited on
Commit
c154065
·
verified ·
1 Parent(s): f032383

Create config.py

Browse files
Files changed (1) hide show
  1. config.py +100 -0
config.py ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ PROVIDER_DISPLAY_NAMES = {
2
+ "openai": "OpenAI",
3
+ "azure_openai": "Azure OpenAI",
4
+ "anthropic": "Anthropic",
5
+ "deepseek": "DeepSeek",
6
+ "google": "Google",
7
+ "alibaba": "Alibaba",
8
+ "moonshot": "MoonShot",
9
+ "unbound": "Unbound AI",
10
+ "ibm": "IBM",
11
+ "grok": "Grok",
12
+ }
13
+
14
+ # Predefined model names for common providers
15
+ model_names = {
16
+ "anthropic": ["claude-3-5-sonnet-20241022", "claude-3-5-sonnet-20240620", "claude-3-opus-20240229"],
17
+ "openai": ["gpt-4o", "gpt-4", "gpt-3.5-turbo", "o3-mini","gpt-oss-20b:free"],
18
+ "deepseek": ["deepseek-chat", "deepseek-reasoner"],
19
+ "google": ["gemini-2.0-flash", "gemini-2.0-flash-thinking-exp", "gemini-1.5-flash-latest",
20
+ "gemini-1.5-flash-8b-latest", "gemini-2.0-flash-thinking-exp-01-21", "gemini-2.0-pro-exp-02-05",
21
+ "gemini-2.5-pro-preview-03-25", "gemini-2.5-flash-preview-04-17"],
22
+ "ollama": ["qwen2.5:7b", "qwen2.5:14b", "qwen2.5:32b", "qwen2.5-coder:14b", "qwen2.5-coder:32b", "llama2:7b",
23
+ "deepseek-r1:14b", "deepseek-r1:32b"],
24
+ "azure_openai": ["gpt-4o", "gpt-4", "gpt-3.5-turbo"],
25
+ "mistral": ["pixtral-large-latest", "mistral-large-latest", "mistral-small-latest", "ministral-8b-latest"],
26
+ "alibaba": ["qwen-plus", "qwen-max", "qwen-vl-max", "qwen-vl-plus", "qwen-turbo", "qwen-long"],
27
+ "moonshot": ["moonshot-v1-32k-vision-preview", "moonshot-v1-8k-vision-preview"],
28
+ "unbound": ["gemini-2.0-flash", "gpt-4o-mini", "gpt-4o", "gpt-4.5-preview"],
29
+ "grok": [
30
+ "grok-3",
31
+ "grok-3-fast",
32
+ "grok-3-mini",
33
+ "grok-3-mini-fast",
34
+ "grok-2-vision",
35
+ "grok-2-image",
36
+ "grok-2",
37
+ ],
38
+ "siliconflow": [
39
+ "deepseek-ai/DeepSeek-R1",
40
+ "deepseek-ai/DeepSeek-V3",
41
+ "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B",
42
+ "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B",
43
+ "deepseek-ai/DeepSeek-R1-Distill-Qwen-7B",
44
+ "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
45
+ "deepseek-ai/DeepSeek-V2.5",
46
+ "deepseek-ai/deepseek-vl2",
47
+ "Qwen/Qwen2.5-72B-Instruct-128K",
48
+ "Qwen/Qwen2.5-72B-Instruct",
49
+ "Qwen/Qwen2.5-32B-Instruct",
50
+ "Qwen/Qwen2.5-14B-Instruct",
51
+ "Qwen/Qwen2.5-7B-Instruct",
52
+ "Qwen/Qwen2.5-Coder-32B-Instruct",
53
+ "Qwen/Qwen2.5-Coder-7B-Instruct",
54
+ "Qwen/Qwen2-7B-Instruct",
55
+ "Qwen/Qwen2-1.5B-Instruct",
56
+ "Qwen/QwQ-32B-Preview",
57
+ "Qwen/Qwen2-VL-72B-Instruct",
58
+ "Qwen/Qwen2.5-VL-32B-Instruct",
59
+ "Qwen/Qwen2.5-VL-72B-Instruct",
60
+ "TeleAI/TeleChat2",
61
+ "THUDM/glm-4-9b-chat",
62
+ "Vendor-A/Qwen/Qwen2.5-72B-Instruct",
63
+ "internlm/internlm2_5-7b-chat",
64
+ "internlm/internlm2_5-20b-chat",
65
+ "Pro/Qwen/Qwen2.5-7B-Instruct",
66
+ "Pro/Qwen/Qwen2-7B-Instruct",
67
+ "Pro/Qwen/Qwen2-1.5B-Instruct",
68
+ "Pro/THUDM/chatglm3-6b",
69
+ "Pro/THUDM/glm-4-9b-chat",
70
+ ],
71
+ "ibm": ["ibm/granite-vision-3.1-2b-preview", "meta-llama/llama-4-maverick-17b-128e-instruct-fp8",
72
+ "meta-llama/llama-3-2-90b-vision-instruct"],
73
+ "modelscope":[
74
+ "Qwen/Qwen2.5-Coder-32B-Instruct",
75
+ "Qwen/Qwen2.5-Coder-14B-Instruct",
76
+ "Qwen/Qwen2.5-Coder-7B-Instruct",
77
+ "Qwen/Qwen2.5-72B-Instruct",
78
+ "Qwen/Qwen2.5-32B-Instruct",
79
+ "Qwen/Qwen2.5-14B-Instruct",
80
+ "Qwen/Qwen2.5-7B-Instruct",
81
+ "Qwen/QwQ-32B-Preview",
82
+ "Qwen/Qwen2.5-VL-3B-Instruct",
83
+ "Qwen/Qwen2.5-VL-7B-Instruct",
84
+ "Qwen/Qwen2.5-VL-32B-Instruct",
85
+ "Qwen/Qwen2.5-VL-72B-Instruct",
86
+ "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B",
87
+ "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B",
88
+ "deepseek-ai/DeepSeek-R1-Distill-Qwen-7B",
89
+ "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
90
+ "deepseek-ai/DeepSeek-R1",
91
+ "deepseek-ai/DeepSeek-V3",
92
+ "Qwen/Qwen3-1.7B",
93
+ "Qwen/Qwen3-4B",
94
+ "Qwen/Qwen3-8B",
95
+ "Qwen/Qwen3-14B",
96
+ "Qwen/Qwen3-30B-A3B",
97
+ "Qwen/Qwen3-32B",
98
+ "Qwen/Qwen3-235B-A22B",
99
+ ],
100
+ }