MarcusAureliusAntoninus commited on
Commit
5e1ed0d
·
verified ·
1 Parent(s): fcf7286

Update librechat.yaml

Browse files
Files changed (1) hide show
  1. librechat.yaml +62 -398
librechat.yaml CHANGED
@@ -1,267 +1,70 @@
1
- version: 1.1.4
 
2
 
 
3
  cache: true
4
 
5
- interface:
6
- privacyPolicy:
7
- externalUrl: 'https://librechat.ai/privacy-policy'
8
- openNewTab: true
9
- termsOfService:
10
- externalUrl: 'https://librechat.ai/tos'
11
- openNewTab: true
12
-
13
- registration:
14
- socialLogins: ["discord", "facebook", "github", "google", "openid"]
15
-
16
  endpoints:
17
  custom:
18
- #cohere
19
- - name: "cohere"
20
- apiKey: "user_provided"
21
- baseURL: "https://api.cohere.ai/v1"
22
- models:
23
- default: ["command-r","command-r-plus","command-r-08-24","command-r-plus-08-24"]
24
- fetch: true
25
- modelDisplayLabel: "cohere"
26
- titleModel: "command"
27
- dropParams: ["stop", "user", "frequency_penalty", "presence_penalty", "temperature", "top_p"]
28
-
29
- #groq
30
- - name: "groq"
31
- apiKey: "user_provided"
32
- baseURL: "https://api.groq.com/openai/v1/"
33
- models:
34
- default: [
35
- "llama2-70b-4096",
36
- "llama3-70b-8192",
37
- "llama3.1-70b-versatile",
38
- "llama3-8b-8192",
39
- "mixtral-8x7b-32768",
40
- "gemma-7b-it",
41
- ]
42
- fetch: true
43
- titleConvo: true
44
- titleModel: "mixtral-8x7b-32768"
45
- modelDisplayLabel: "groq"
46
-
47
  # Mistral AI API
48
- - name: "Mistral"
49
- apiKey: "user_provided"
 
 
50
  baseURL: "https://api.mistral.ai/v1"
 
 
51
  models:
52
- default: [
53
- "mistral-large-latest",
54
- "mistral-small",
55
- "mistral-medium"
56
- ]
57
- fetch: false
58
- titleConvo: true
59
- titleMethod: "completion"
60
- titleModel: "mistral-large-latest"
 
 
 
 
 
 
 
 
61
  summarize: false
62
- summaryModel: "mistral-tiny"
 
 
 
 
63
  forcePrompt: false
64
- modelDisplayLabel: "Mistral"
65
- dropParams: ["stop", "user", "frequency_penalty", "presence_penalty"]
66
 
67
- # OpenRouter.ai
 
 
 
 
 
 
 
 
 
 
 
 
 
 
68
  - name: "OpenRouter"
69
- apiKey: "user_provided"
 
 
 
70
  baseURL: "https://openrouter.ai/api/v1"
71
  models:
72
- "default": [
73
- "openrouter/auto",
74
- "---FREE---",
75
- "google/gemma-7b-it:free",
76
- "gryphe/mythomist-7b:free",
77
- "huggingfaceh4/zephyr-7b-beta:free",
78
- "meta-llama/llama-3-8b-instruct:free",
79
- "microsoft/phi-3-medium-128k-instruct:free",
80
- "microsoft/phi-3-mini-128k-instruct:free",
81
- "mistralai/mistral-7b-instruct:free",
82
- "nousresearch/nous-capybara-7b:free",
83
- "openchat/openchat-7b:free",
84
- "undi95/toppy-m-7b:free",
85
- "---NITRO---",
86
- "google/gemma-7b-it:nitro",
87
- "gryphe/mythomax-l2-13b:nitro",
88
- "meta-llama/llama-2-70b-chat:nitro",
89
- "meta-llama/llama-3-70b-instruct:nitro",
90
- "meta-llama/llama-3-8b-instruct:nitro",
91
- "mistralai/mistral-7b-instruct:nitro",
92
- "mistralai/mixtral-8x7b-instruct:nitro",
93
- "undi95/toppy-m-7b:nitro",
94
- "---BETA---",
95
- "anthropic/claude-2.0:beta",
96
- "anthropic/claude-2.1:beta",
97
- "anthropic/claude-2:beta",
98
- "anthropic/claude-3-haiku:beta",
99
- "anthropic/claude-3-opus:beta",
100
- "anthropic/claude-3-sonnet:beta",
101
- "anthropic/claude-3.5-sonnet:beta",
102
- "anthropic/claude-instant-1:beta",
103
- "---EXTENDED---",
104
- "gryphe/mythomax-l2-13b:extended",
105
- "meta-llama/llama-3-8b-instruct:extended",
106
- "neversleep/llama-3-lumimaid-8b:extended",
107
- "undi95/remm-slerp-l2-13b:extended",
108
- "---01-AI---",
109
- "01-ai/yi-34b",
110
- "01-ai/yi-34b-chat",
111
- "01-ai/yi-6b",
112
- "---ANTHROPIC---",
113
- "anthropic/claude-1",
114
- "anthropic/claude-1.2",
115
- "anthropic/claude-2",
116
- "anthropic/claude-2.0",
117
- "anthropic/claude-2.1",
118
- "anthropic/claude-3-haiku",
119
- "anthropic/claude-3-opus",
120
- "anthropic/claude-3-sonnet",
121
- "anthropic/claude-3.5-sonnet",
122
- "anthropic/claude-instant-1",
123
- "anthropic/claude-instant-1.0",
124
- "anthropic/claude-instant-1.1",
125
- "---COHERE---",
126
- "cohere/command",
127
- "cohere/command-r",
128
- "cohere/command-r-plus",
129
- "---GOOGLE---",
130
- "google/gemini-flash-1.5",
131
- "google/gemini-pro",
132
- "google/gemini-pro-1.5",
133
- "google/gemini-pro-vision",
134
- "google/gemma-7b-it",
135
- "google/palm-2-chat-bison",
136
- "google/palm-2-chat-bison-32k",
137
- "google/palm-2-codechat-bison",
138
- "google/palm-2-codechat-bison-32k",
139
- "---META-LLAMA---",
140
- "meta-llama/codellama-34b-instruct",
141
- "meta-llama/codellama-70b-instruct",
142
- "meta-llama/llama-2-13b-chat",
143
- "meta-llama/llama-2-70b-chat",
144
- "meta-llama/llama-3-70b",
145
- "meta-llama/llama-3-70b-instruct",
146
- "meta-llama/llama-3-8b",
147
- "meta-llama/llama-3-8b-instruct",
148
- "meta-llama/llama-guard-2-8b",
149
- "---MICROSOFT---",
150
- "microsoft/phi-3-medium-128k-instruct",
151
- "microsoft/phi-3-medium-4k-instruct",
152
- "microsoft/phi-3-mini-128k-instruct",
153
- "microsoft/wizardlm-2-7b",
154
- "microsoft/wizardlm-2-8x22b",
155
- "---MISTRALAI---",
156
- "mistralai/mistral-7b-instruct",
157
- "mistralai/mistral-7b-instruct-v0.1",
158
- "mistralai/mistral-7b-instruct-v0.2",
159
- "mistralai/mistral-7b-instruct-v0.3",
160
- "mistralai/mistral-large",
161
- "mistralai/mistral-medium",
162
- "mistralai/mistral-small",
163
- "mistralai/mistral-tiny",
164
- "mistralai/mixtral-8x22b",
165
- "mistralai/mixtral-8x22b-instruct",
166
- "mistralai/mixtral-8x7b",
167
- "mistralai/mixtral-8x7b-instruct",
168
- "---NEVERSLEEP---",
169
- "neversleep/llama-3-lumimaid-70b",
170
- "neversleep/llama-3-lumimaid-8b",
171
- "neversleep/noromaid-20b",
172
- "neversleep/noromaid-mixtral-8x7b-instruct",
173
- "---NOUSRESEARCH---",
174
- "nousresearch/hermes-2-pro-llama-3-8b",
175
- "nousresearch/nous-capybara-34b",
176
- "nousresearch/nous-capybara-7b",
177
- "nousresearch/nous-hermes-2-mistral-7b-dpo",
178
- "nousresearch/nous-hermes-2-mixtral-8x7b-dpo",
179
- "nousresearch/nous-hermes-2-mixtral-8x7b-sft",
180
- "nousresearch/nous-hermes-2-vision-7b",
181
- "nousresearch/nous-hermes-llama2-13b",
182
- "nousresearch/nous-hermes-llama2-70b",
183
- "nousresearch/nous-hermes-yi-34b",
184
- "---OPENAI---",
185
- "openai/gpt-3.5-turbo",
186
- "openai/gpt-3.5-turbo-0125",
187
- "openai/gpt-3.5-turbo-0301",
188
- "openai/gpt-3.5-turbo-0613",
189
- "openai/gpt-3.5-turbo-1106",
190
- "openai/gpt-3.5-turbo-16k",
191
- "openai/gpt-3.5-turbo-instruct",
192
- "openai/gpt-4",
193
- "openai/gpt-4-0314",
194
- "openai/gpt-4-1106-preview",
195
- "openai/gpt-4-32k",
196
- "openai/gpt-4-32k-0314",
197
- "openai/gpt-4-turbo",
198
- "openai/gpt-4-turbo-preview",
199
- "openai/gpt-4-vision-preview",
200
- "openai/gpt-4o",
201
- "openai/gpt-4o-2024-05-13",
202
- "---PERPLEXITY---",
203
- "perplexity/llama-3-sonar-large-32k-chat",
204
- "perplexity/llama-3-sonar-large-32k-online",
205
- "perplexity/llama-3-sonar-small-32k-chat",
206
- "perplexity/llama-3-sonar-small-32k-online",
207
- "---QWEN---",
208
- "qwen/qwen-110b-chat",
209
- "qwen/qwen-14b-chat",
210
- "qwen/qwen-2-72b-instruct",
211
- "qwen/qwen-32b-chat",
212
- "qwen/qwen-4b-chat",
213
- "qwen/qwen-72b-chat",
214
- "qwen/qwen-7b-chat",
215
- "---OTHERS---",
216
- "ai21/jamba-instruct",
217
- "allenai/olmo-7b-instruct",
218
- "alpindale/goliath-120b",
219
- "austism/chronos-hermes-13b",
220
- "bigcode/starcoder2-15b-instruct",
221
- "cognitivecomputations/dolphin-mixtral-8x22b",
222
- "cognitivecomputations/dolphin-mixtral-8x7b",
223
- "databricks/dbrx-instruct",
224
- "deepseek/deepseek-chat",
225
- "deepseek/deepseek-coder",
226
- "fireworks/firellava-13b",
227
- "gryphe/mythomax-l2-13b",
228
- "gryphe/mythomist-7b",
229
- "huggingfaceh4/zephyr-7b-beta",
230
- "huggingfaceh4/zephyr-orpo-141b-a35b",
231
- "intel/neural-chat-7b",
232
- "jebcarter/psyfighter-13b",
233
- "jondurbin/airoboros-l2-70b",
234
- "jondurbin/bagel-34b",
235
- "koboldai/psyfighter-13b-2",
236
- "liuhaotian/llava-13b",
237
- "liuhaotian/llava-yi-34b",
238
- "lizpreciatior/lzlv-70b-fp16-hf",
239
- "lynn/soliloquy-l3",
240
- "mancer/weaver",
241
- "nvidia/nemotron-4-340b-instruct",
242
- "open-orca/mistral-7b-openorca",
243
- "openchat/openchat-7b",
244
- "openchat/openchat-8b",
245
- "openrouter/flavor-of-the-week",
246
- "phind/phind-codellama-34b",
247
- "pygmalionai/mythalion-13b",
248
- "recursal/eagle-7b",
249
- "recursal/rwkv-5-3b-ai-town",
250
- "rwkv/rwkv-5-world-3b",
251
- "sao10k/fimbulvetr-11b-v2",
252
- "sao10k/l3-euryale-70b",
253
- "snowflake/snowflake-arctic-instruct",
254
- "sophosympatheia/midnight-rose-70b",
255
- "teknium/openhermes-2-mistral-7b",
256
- "teknium/openhermes-2.5-mistral-7b",
257
- "togethercomputer/stripedhyena-hessian-7b",
258
- "togethercomputer/stripedhyena-nous-7b",
259
- "undi95/remm-slerp-l2-13b",
260
- "undi95/toppy-m-7b",
261
- "xwin-lm/xwin-lm-70b"
262
- ]
263
  fetch: true
264
- dropParams: ["stop"]
265
  titleConvo: true
266
  titleModel: "gpt-3.5-turbo"
267
  summarize: false
@@ -269,160 +72,21 @@ endpoints:
269
  forcePrompt: false
270
  modelDisplayLabel: "OpenRouter"
271
 
272
- # # OpenRouter.ai - Perplexity
273
- # - name: "OpenRouter-Perplexity"
274
- # apiKey: "user_provided"
275
- # baseURL: "https://openrouter.ai/api/v1"
276
- # models:
277
- # "default": [
278
- # "perplexity/pplx-7b-chat",
279
- # "perplexity/pplx-70b-chat",
280
- # "perplexity/pplx-7b-online",
281
- # "perplexity/pplx-70b-online",
282
- # "perplexity/sonar-medium-online",
283
- # ]
284
- # fetch: false
285
- # titleConvo: true
286
- # titleModel: "perplexity/pplx-7b-chat"
287
- # summarize: false
288
- # summaryModel: "perplexity/pplx-7b-chat"
289
- # forcePrompt: false
290
- # dropParams: ["stop"]
291
- # modelDisplayLabel: "Perplexity"
292
- # iconURL: "https://raw.githubusercontent.com/fuegovic/lc-config-yaml/main/icons/perplexityai.png"
293
-
294
- # Preplexity
295
- - name: "Perplexity"
296
  apiKey: "user_provided"
297
- baseURL: "https://api.perplexity.ai/"
298
  models:
299
- default: [
300
- "llama-3-sonar-small-32k-chat",
301
- "llama-3-sonar-small-32k-online",
302
- "llama-3-sonar-large-32k-chat",
303
- "llama-3-sonar-large-32k-online",
304
- "llama-3-8b-instruct",
305
- "llama-3-70b-instruct",
306
- "sonar-medium-online"
307
- ]
308
- fetch: false # fetching list of models is not supported
309
- titleConvo: true
310
- titleModel: "sonar-medium-chat"
311
- summarize: false
312
- summaryModel: "sonar-medium-chat"
313
- forcePrompt: false
314
- dropParams: ["stop", "frequency_penalty"]
315
- modelDisplayLabel: "Perplexity"
316
-
317
- # ShuttleAI API
318
- - name: "ShuttleAI"
319
- apiKey: "user_provided"
320
- baseURL: "https://api.shuttleai.app/v1"
321
- models:
322
- default: [
323
- "shuttle-2-turbo",
324
- "shuttle-turbo",
325
- "shuttleai/shuttle-2.5",
326
- "shuttle-2.5",
327
- "im-also-a-good-gpt2-chatbot",
328
- "gpt-4-turbo-2024-04-09",
329
- "gpt-4-turbo",
330
- "gpt-4-0125-preview",
331
- "gpt-4-turbo-preview",
332
- "gpt-4-1106-preview",
333
- "gpt-4-1106-vision-preview",
334
- "gpt-4-vision-preview",
335
- "gpt-4-0613",
336
- "gpt-4",
337
- "gpt-4-bing",
338
- "gpt-4-turbo-bing",
339
- "gpt-4-32k-0613",
340
- "gpt-4-32k",
341
- "gpt-3.5-turbo-0125",
342
- "gpt-3.5-turbo",
343
- "gpt-3.5-turbo-1106",
344
- "claude-3-opus-20240229",
345
- "claude-3-opus",
346
- "claude-3-sonnet-20240229",
347
- "claude-3-sonnet",
348
- "claude-3-haiku-20240307",
349
- "claude-3-haiku",
350
- "claude-2.1",
351
- "claude-2.0",
352
- "claude-2",
353
- "claude-instant-1.2",
354
- "claude-instant-1.1",
355
- "claude-instant-1.0",
356
- "claude-instant",
357
- "meta-llama-3-70b-instruct",
358
- "llama-3-70b-instruct",
359
- "meta-llama-3-8b-instruct",
360
- "llama-3-8b-instruct",
361
- "llama-3-sonar-large-32k-online",
362
- "llama-3-sonar-small-32k-online",
363
- "llama-3-sonar-large-32k-chat",
364
- "llama-3-sonar-small-32k-chat",
365
- "llama-3.1-sonar-small-128k-online",
366
- "blackbox",
367
- "blackbox-code",
368
- "wizardlm-2-8x22b",
369
- "wizardlm-2-70b",
370
- "dolphin-2.6-mixtral-8x7b",
371
- "dolphin-mixtral-8x7b",
372
- "mistral-large",
373
- "mistral-next",
374
- "mistral-medium",
375
- "mistral-small",
376
- "mistral-tiny",
377
- "mixtral-8x7b-instruct-v0.1",
378
- "mixtral-8x7b-instruct",
379
- "mixtral-8x22b-instruct-v0.1",
380
- "mixtral-8x22b-instruct",
381
- "mistral-7b-instruct-v0.2",
382
- "mistral-7b-instruct-2",
383
- "mistral-7b-instruct-v0.1",
384
- "mistral-7b-instruct",
385
- "nous-hermes-2-mixtral-8x7b",
386
- "gemini-1.5-pro-latest",
387
- "gemini-1.5-pro",
388
- "gemini-1.0-pro-latest",
389
- "gemini-1.0-pro",
390
- "gemini-pro",
391
- "gemini-1.0-pro-vision",
392
- "gemini-pro-vision",
393
- "lzlv-70b",
394
- "duckduckgo/search-ddg",
395
- "google/search-google"
396
- ]
397
- fetch: false
398
- titleConvo: true
399
- titleMethod: "completion"
400
- titleModel: "shuttle-2-turbo"
401
- summarize: false
402
- summaryModel: "shuttle-2-turbo"
403
- forcePrompt: false
404
- dropParams: ["user", "frequency_penalty", "presence_penalty", "repetition_penalty"]
405
- modelDisplayLabel: "ShuttleAI"
406
- # iconURL: "https://raw.githubusercontent.com/fuegovic/lc-config-yaml/main/icons/ShuttleAI.png"
407
-
408
- # Whore
409
-
410
- # Whore
411
- - name: "Whore"
412
- apiKey: "${WHORE_API_KEY}"
413
- baseURL: ""
414
- models:
415
- default: [
416
- "gpt-4",
417
- "gpt-4-32k",
418
- "gpt-4-1106-preview",
419
- "gpt-4-0125-preview",
420
- ]
421
  fetch: true
422
  titleConvo: true
423
  titleModel: "gpt-3.5-turbo"
424
  summarize: false
425
  summaryModel: "gpt-3.5-turbo"
426
  forcePrompt: false
427
- dropParams: ["stream"]
428
- iconURL: "https://raw.githubusercontent.com/fuegovic/lc-config-yaml/main/icons/zuki.png"
 
 
 
1
+ # Configuration version (required)
2
+ version: 1.0.0
3
 
4
+ # Cache settings: Set to true to enable caching
5
  cache: true
6
 
7
+ # Definition of custom endpoints
 
 
 
 
 
 
 
 
 
 
8
  endpoints:
9
  custom:
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  # Mistral AI API
11
+ - name: "Mistral" # Unique name for the endpoint
12
+ # For `apiKey` and `baseURL`, you can use environment variables that you define.
13
+ # recommended environment variables:
14
+ apiKey: "${MISTRAL_API_KEY}"
15
  baseURL: "https://api.mistral.ai/v1"
16
+
17
+ # Models configuration
18
  models:
19
+ # List of default models to use. At least one value is required.
20
+ default: ["mistral-tiny", "mistral-small", "mistral-medium"]
21
+ # Fetch option: Set to true to fetch models from API.
22
+ fetch: true # Defaults to false.
23
+
24
+ # Optional configurations
25
+
26
+ # Title Conversation setting
27
+ titleConvo: true # Set to true to enable title conversation
28
+
29
+ # Title Method: Choose between "completion" or "functions".
30
+ titleMethod: "completion" # Defaults to "completion" if omitted.
31
+
32
+ # Title Model: Specify the model to use for titles.
33
+ titleModel: "mistral-tiny" # Defaults to "gpt-3.5-turbo" if omitted.
34
+
35
+ # Summarize setting: Set to true to enable summarization.
36
  summarize: false
37
+
38
+ # Summary Model: Specify the model to use if summarization is enabled.
39
+ summaryModel: "mistral-tiny" # Defaults to "gpt-3.5-turbo" if omitted.
40
+
41
+ # Force Prompt setting: If true, sends a `prompt` parameter instead of `messages`.
42
  forcePrompt: false
 
 
43
 
44
+ # The label displayed for the AI model in messages.
45
+ modelDisplayLabel: "Mistral" # Default is "AI" when not set.
46
+
47
+ # Add additional parameters to the request. Default params will be overwritten.
48
+ addParams:
49
+ safe_mode: true # This field is specific to Mistral AI: https://docs.mistral.ai/api/
50
+
51
+ # Drop Default params parameters from the request. See default params in guide linked below.
52
+ dropParams: ["stop", "temperature", "top_p"]
53
+ # - stop # dropped since it's not recognized by Mistral AI API
54
+ # `temperature` and `top_p` are removed to allow Mistral AI API defaults to be used:
55
+ # - temperature
56
+ # - top_p
57
+
58
+ # OpenRouter.ai Example
59
  - name: "OpenRouter"
60
+ # For `apiKey` and `baseURL`, you can use environment variables that you define.
61
+ # recommended environment variables:
62
+ # Known issue: you should not use `OPENROUTER_API_KEY` as it will then override the `openAI` endpoint to use OpenRouter as well.
63
+ apiKey: "${OPENROUTER_KEY}"
64
  baseURL: "https://openrouter.ai/api/v1"
65
  models:
66
+ default: ["nousresearch/nous-capybara-7b:free", "mistralai/mistral-7b-instruct:free", "huggingfaceh4/zephyr-7b-beta:free", "openchat/openchat-7b:free", "gryphe/mythomist-7b:free", "undi95/toppy-m-7b:free", "openrouter/cinematika-7b:free", "openrouter/auto", "nousresearch/nous-capybara-7b", "mistralai/mistral-7b-instruct", "huggingfaceh4/zephyr-7b-beta", "openchat/openchat-7b", "gryphe/mythomist-7b", "openrouter/cinematika-7b", "rwkv/rwkv-5-world-3b", "recursal/rwkv-5-3b-ai-town", "jondurbin/bagel-34b", "jebcarter/psyfighter-13b", "koboldai/psyfighter-13b-2", "neversleep/noromaid-mixtral-8x7b-instruct", "nousresearch/nous-hermes-llama2-13b", "meta-llama/codellama-34b-instruct", "phind/phind-codellama-34b", "intel/neural-chat-7b", "nousresearch/nous-hermes-2-mixtral-8x7b-dpo", "nousresearch/nous-hermes-2-mixtral-8x7b-sft", "haotian-liu/llava-13b", "nousresearch/nous-hermes-2-vision-7b", "meta-llama/llama-2-13b-chat", "gryphe/mythomax-l2-13b", "nousresearch/nous-hermes-llama2-70b", "teknium/openhermes-2-mistral-7b", "teknium/openhermes-2.5-mistral-7b", "undi95/remm-slerp-l2-13b", "undi95/toppy-m-7b", "01-ai/yi-34b-chat", "01-ai/yi-34b", "01-ai/yi-6b", "togethercomputer/stripedhyena-nous-7b", "togethercomputer/stripedhyena-hessian-7b", "mistralai/mixtral-8x7b", "nousresearch/nous-hermes-yi-34b", "open-orca/mistral-7b-openorca", "openai/gpt-3.5-turbo", "openai/gpt-3.5-turbo-1106", "openai/gpt-3.5-turbo-16k", "openai/gpt-4-1106-preview", "openai/gpt-4", "openai/gpt-4-32k", "openai/gpt-4-vision-preview", "openai/gpt-3.5-turbo-instruct", "google/palm-2-chat-bison", "google/palm-2-codechat-bison", "google/palm-2-chat-bison-32k", "google/palm-2-codechat-bison-32k", "google/gemini-pro", "google/gemini-pro-vision", "perplexity/pplx-70b-online", "perplexity/pplx-7b-online", "perplexity/pplx-7b-chat", "perplexity/pplx-70b-chat", "meta-llama/llama-2-70b-chat", "nousresearch/nous-capybara-34b", "jondurbin/airoboros-l2-70b", "austism/chronos-hermes-13b", "migtissera/synthia-70b", "pygmalionai/mythalion-13b", "undi95/remm-slerp-l2-13b-6k", "xwin-lm/xwin-lm-70b", "gryphe/mythomax-l2-13b-8k", "alpindale/goliath-120b ", "lizpreciatior/lzlv-70b-fp16-hf", "neversleep/noromaid-20b", "mistralai/mixtral-8x7b-instruct", "cognitivecomputations/dolphin-mixtral-8x7b", "anthropic/claude-2", "anthropic/claude-2.0", "anthropic/claude-instant-v1", "mancer/weaver", "mistralai/mistral-tiny", "mistralai/mistral-small", "mistralai/mistral-medium"]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
67
  fetch: true
 
68
  titleConvo: true
69
  titleModel: "gpt-3.5-turbo"
70
  summarize: false
 
72
  forcePrompt: false
73
  modelDisplayLabel: "OpenRouter"
74
 
75
+ - name: "Reverse Proxy"
76
+ # For `apiKey` and `baseURL`, you can use environment variables that you define.
77
+ # recommended environment variables:
78
+ # Known issue: you should not use `OPENROUTER_API_KEY` as it will then override the `openAI` endpoint to use OpenRouter as well.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
79
  apiKey: "user_provided"
80
+ baseURL: "user_provided"
81
  models:
82
+ default: ["gpt-3.5-turbo"]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
83
  fetch: true
84
  titleConvo: true
85
  titleModel: "gpt-3.5-turbo"
86
  summarize: false
87
  summaryModel: "gpt-3.5-turbo"
88
  forcePrompt: false
89
+ modelDisplayLabel: "AI"
90
+
91
+ # See the Custom Configuration Guide for more information:
92
+ # https://docs.librechat.ai/install/configuration/custom_config.html