Update librechat.yaml
Browse files- librechat.yaml +17 -16
librechat.yaml
CHANGED
|
@@ -1,4 +1,4 @@
|
|
| 1 |
-
version: 1.2.
|
| 2 |
|
| 3 |
cache: true
|
| 4 |
|
|
@@ -32,6 +32,18 @@ modelSpecs:
|
|
| 32 |
topK: 40
|
| 33 |
topP: 0.95
|
| 34 |
maxOutputTokens: 8192
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 35 |
- name: "o3-mini"
|
| 36 |
label: "o3-mini - 12"
|
| 37 |
iconURL: "https://cdn.jsdelivr.net/gh/Guru-25/Nothing/librechat/chatgpt.png"
|
|
@@ -39,7 +51,7 @@ modelSpecs:
|
|
| 39 |
modelLabel: "o3-mini"
|
| 40 |
endpoint: "Github Models"
|
| 41 |
model: "o3-mini"
|
| 42 |
-
|
| 43 |
reasoning_effort: "high"
|
| 44 |
- name: "o1"
|
| 45 |
label: "o1 - 8*"
|
|
@@ -48,7 +60,7 @@ modelSpecs:
|
|
| 48 |
modelLabel: "o1"
|
| 49 |
endpoint: "Github Models"
|
| 50 |
model: "o1"
|
| 51 |
-
|
| 52 |
reasoning_effort: "high"
|
| 53 |
- name: "deepseek-r1"
|
| 54 |
label: "DeepSeek R1 - 8"
|
|
@@ -58,18 +70,6 @@ modelSpecs:
|
|
| 58 |
endpoint: "Github Models"
|
| 59 |
model: "deepseek-r1"
|
| 60 |
max_tokens: 2048
|
| 61 |
-
- name: "deepseek-v3"
|
| 62 |
-
label: "DeepSeek V3 - 50*"
|
| 63 |
-
iconURL: "https://cdn.jsdelivr.net/gh/Guru-25/Nothing/librechat/deepseek.svg"
|
| 64 |
-
preset:
|
| 65 |
-
modelLabel: "DeepSeek V3"
|
| 66 |
-
endpoint: "Github Models"
|
| 67 |
-
model: "deepseek-v3"
|
| 68 |
-
max_completion_tokens: 2048
|
| 69 |
-
temperature: 0.8
|
| 70 |
-
top_p: 0.1
|
| 71 |
-
presence_penalty: 0
|
| 72 |
-
frequency_penalty: 0
|
| 73 |
- name: "deepseek-r1-openrouter"
|
| 74 |
label: "DeepSeek R1 (OpenRouter) - 200*"
|
| 75 |
iconURL: "https://cdn.jsdelivr.net/gh/Guru-25/Nothing/librechat/deepseek.svg"
|
|
@@ -123,8 +123,9 @@ endpoints:
|
|
| 123 |
models:
|
| 124 |
default:
|
| 125 |
- deepseek/deepseek-r1:free
|
|
|
|
| 126 |
titleConvo: true
|
| 127 |
-
titleModel: "google/gemini-2.0-
|
| 128 |
|
| 129 |
# groq
|
| 130 |
# Model list: https://console.groq.com/settings/limits
|
|
|
|
| 1 |
+
version: 1.2.3
|
| 2 |
|
| 3 |
cache: true
|
| 4 |
|
|
|
|
| 32 |
topK: 40
|
| 33 |
topP: 0.95
|
| 34 |
maxOutputTokens: 8192
|
| 35 |
+
- name: "deepseek-v3"
|
| 36 |
+
label: "DeepSeek V3 - 50"
|
| 37 |
+
iconURL: "https://cdn.jsdelivr.net/gh/Guru-25/Nothing/librechat/deepseek.svg"
|
| 38 |
+
preset:
|
| 39 |
+
modelLabel: "DeepSeek V3"
|
| 40 |
+
endpoint: "Github Models"
|
| 41 |
+
model: "deepseek-v3"
|
| 42 |
+
max_tokens: 2048
|
| 43 |
+
temperature: 0.8
|
| 44 |
+
top_p: 0.1
|
| 45 |
+
presence_penalty: 0
|
| 46 |
+
frequency_penalty: 0
|
| 47 |
- name: "o3-mini"
|
| 48 |
label: "o3-mini - 12"
|
| 49 |
iconURL: "https://cdn.jsdelivr.net/gh/Guru-25/Nothing/librechat/chatgpt.png"
|
|
|
|
| 51 |
modelLabel: "o3-mini"
|
| 52 |
endpoint: "Github Models"
|
| 53 |
model: "o3-mini"
|
| 54 |
+
max_tokens: 100000
|
| 55 |
reasoning_effort: "high"
|
| 56 |
- name: "o1"
|
| 57 |
label: "o1 - 8*"
|
|
|
|
| 60 |
modelLabel: "o1"
|
| 61 |
endpoint: "Github Models"
|
| 62 |
model: "o1"
|
| 63 |
+
max_tokens: 40000
|
| 64 |
reasoning_effort: "high"
|
| 65 |
- name: "deepseek-r1"
|
| 66 |
label: "DeepSeek R1 - 8"
|
|
|
|
| 70 |
endpoint: "Github Models"
|
| 71 |
model: "deepseek-r1"
|
| 72 |
max_tokens: 2048
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 73 |
- name: "deepseek-r1-openrouter"
|
| 74 |
label: "DeepSeek R1 (OpenRouter) - 200*"
|
| 75 |
iconURL: "https://cdn.jsdelivr.net/gh/Guru-25/Nothing/librechat/deepseek.svg"
|
|
|
|
| 123 |
models:
|
| 124 |
default:
|
| 125 |
- deepseek/deepseek-r1:free
|
| 126 |
+
- google/gemini-2.0-flash-exp:free
|
| 127 |
titleConvo: true
|
| 128 |
+
titleModel: "google/gemini-2.0-flash-exp:free"
|
| 129 |
|
| 130 |
# groq
|
| 131 |
# Model list: https://console.groq.com/settings/limits
|