Spaces:
Sleeping
Sleeping
Bump, pull in generated configs from personal setup
Browse files- Dockerfile +4 -4
- azure-models.txt +0 -27
- generated/anthropic.yaml +24 -0
- generated/github.yaml +56 -0
- generated/groq.yaml +8 -0
- generated/huggingface.yaml +15 -0
- generated/openai.yaml +32 -0
- generated/perplexity.yaml +12 -0
- start.sh +18 -26
Dockerfile
CHANGED
|
@@ -1,16 +1,16 @@
|
|
| 1 |
# Building from source is too slow, so let's just customize the existing image
|
| 2 |
-
FROM ghcr.io/open-webui/open-webui:git-
|
| 3 |
|
| 4 |
WORKDIR /app
|
| 5 |
|
| 6 |
USER 0:0
|
| 7 |
|
| 8 |
# HACK for huggingface.co iframe
|
| 9 |
-
RUN sed -i "s|samesite=WEBUI_SESSION_COOKIE_SAME_SITE|samesite='none'|g" backend/open_webui/
|
| 10 |
|
| 11 |
-
RUN pip install "litellm[proxy]==1.
|
| 12 |
USER 1000:0
|
| 13 |
|
| 14 |
-
COPY ./
|
| 15 |
COPY ./start.sh /start.sh
|
| 16 |
CMD [ "bash", "/start.sh" ]
|
|
|
|
| 1 |
# Building from source is too slow, so let's just customize the existing image
|
| 2 |
+
FROM ghcr.io/open-webui/open-webui:git-b72150c
|
| 3 |
|
| 4 |
WORKDIR /app
|
| 5 |
|
| 6 |
USER 0:0
|
| 7 |
|
| 8 |
# HACK for huggingface.co iframe
|
| 9 |
+
RUN sed -i "s|samesite=WEBUI_SESSION_COOKIE_SAME_SITE|samesite='none'|g" backend/open_webui/routers/auths.py
|
| 10 |
|
| 11 |
+
RUN pip install "litellm[proxy]==1.59.9" && chown -R 1000:0 /app
|
| 12 |
USER 1000:0
|
| 13 |
|
| 14 |
+
COPY ./generated /assets/
|
| 15 |
COPY ./start.sh /start.sh
|
| 16 |
CMD [ "bash", "/start.sh" ]
|
azure-models.txt
DELETED
|
@@ -1,27 +0,0 @@
|
|
| 1 |
-
AI21-Jamba-Instruct
|
| 2 |
-
Cohere-command-r
|
| 3 |
-
Cohere-command-r-plus
|
| 4 |
-
Cohere-embed-v3-english
|
| 5 |
-
Cohere-embed-v3-multilingual
|
| 6 |
-
Meta-Llama-3-70B-Instruct
|
| 7 |
-
Meta-Llama-3-8B-Instruct
|
| 8 |
-
Meta-Llama-3.1-405B-Instruct
|
| 9 |
-
Meta-Llama-3.1-70B-Instruct
|
| 10 |
-
Meta-Llama-3.1-8B-Instruct
|
| 11 |
-
Mistral-large
|
| 12 |
-
Mistral-large-2407
|
| 13 |
-
Mistral-Nemo
|
| 14 |
-
Mistral-small
|
| 15 |
-
gpt-4o
|
| 16 |
-
gpt-4o-mini
|
| 17 |
-
text-embedding-3-large
|
| 18 |
-
text-embedding-3-small
|
| 19 |
-
Phi-3-medium-128k-instruct
|
| 20 |
-
Phi-3-medium-4k-instruct
|
| 21 |
-
Phi-3-mini-128k-instruct
|
| 22 |
-
Phi-3-mini-4k-instruct
|
| 23 |
-
Phi-3-small-128k-instruct
|
| 24 |
-
Phi-3-small-8k-instruct
|
| 25 |
-
Phi-3.5-mini-instruct
|
| 26 |
-
o1-preview
|
| 27 |
-
o1-mini
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
generated/anthropic.yaml
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
- litellm_params:
|
| 2 |
+
api_key: os.environ/ANTHROPIC_API_KEY
|
| 3 |
+
model: anthropic/claude-3-5-haiku-latest
|
| 4 |
+
model_name: anthropic/claude-3-5-haiku
|
| 5 |
+
- litellm_params:
|
| 6 |
+
api_key: os.environ/ANTHROPIC_API_KEY
|
| 7 |
+
model: anthropic/claude-3-5-sonnet-latest
|
| 8 |
+
model_name: anthropic/claude-3-5-sonnet
|
| 9 |
+
- litellm_params:
|
| 10 |
+
api_key: os.environ/ANTHROPIC_API_KEY
|
| 11 |
+
model: anthropic/claude-3-5-opus-latest
|
| 12 |
+
model_name: anthropic/claude-3-5-opus
|
| 13 |
+
- litellm_params:
|
| 14 |
+
api_key: os.environ/ANTHROPIC_API_KEY
|
| 15 |
+
model: anthropic/claude-3-haiku-latest
|
| 16 |
+
model_name: anthropic/claude-3-haiku
|
| 17 |
+
- litellm_params:
|
| 18 |
+
api_key: os.environ/ANTHROPIC_API_KEY
|
| 19 |
+
model: anthropic/claude-3-sonnet-latest
|
| 20 |
+
model_name: anthropic/claude-3-sonnet
|
| 21 |
+
- litellm_params:
|
| 22 |
+
api_key: os.environ/ANTHROPIC_API_KEY
|
| 23 |
+
model: anthropic/claude-3-opus-latest
|
| 24 |
+
model_name: anthropic/claude-3-opus
|
generated/github.yaml
ADDED
|
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
- litellm_params:
|
| 2 |
+
api_key: os.environ/GITHUB_API_KEY
|
| 3 |
+
model: github/DeepSeek-R1
|
| 4 |
+
model_name: deepseek/r1
|
| 5 |
+
- litellm_params:
|
| 6 |
+
api_key: os.environ/GITHUB_API_KEY
|
| 7 |
+
model: github/gpt-4o
|
| 8 |
+
model_name: openai/gpt-4o
|
| 9 |
+
- litellm_params:
|
| 10 |
+
api_key: os.environ/GITHUB_API_KEY
|
| 11 |
+
model: github/gpt-4o-mini
|
| 12 |
+
model_name: openai/gpt-4o-mini
|
| 13 |
+
- litellm_params:
|
| 14 |
+
api_key: os.environ/GITHUB_API_KEY
|
| 15 |
+
model: github/o1
|
| 16 |
+
model_name: openai/o1
|
| 17 |
+
- litellm_params:
|
| 18 |
+
api_key: os.environ/GITHUB_API_KEY
|
| 19 |
+
model: github/o1-mini
|
| 20 |
+
model_name: openai/o1-mini
|
| 21 |
+
- litellm_params:
|
| 22 |
+
api_key: os.environ/GITHUB_API_KEY
|
| 23 |
+
model: github/text-embedding-3-small
|
| 24 |
+
model_name: openai/text-embedding-3-small
|
| 25 |
+
- litellm_params:
|
| 26 |
+
api_key: os.environ/GITHUB_API_KEY
|
| 27 |
+
model: github/text-embedding-3-large
|
| 28 |
+
model_name: openai/text-embedding-3-large
|
| 29 |
+
- litellm_params:
|
| 30 |
+
api_key: os.environ/GITHUB_API_KEY
|
| 31 |
+
model: github/Mistral-small
|
| 32 |
+
model_name: mistral/small
|
| 33 |
+
- litellm_params:
|
| 34 |
+
api_key: os.environ/GITHUB_API_KEY
|
| 35 |
+
model: github/Mistral-large
|
| 36 |
+
model_name: mistral/large
|
| 37 |
+
- litellm_params:
|
| 38 |
+
api_key: os.environ/GITHUB_API_KEY
|
| 39 |
+
model: github/Mistral-Nemo
|
| 40 |
+
model_name: mistral/nemo
|
| 41 |
+
- litellm_params:
|
| 42 |
+
api_key: os.environ/GITHUB_API_KEY
|
| 43 |
+
model: github/Cohere-command-r
|
| 44 |
+
model_name: cohere/command-r
|
| 45 |
+
- litellm_params:
|
| 46 |
+
api_key: os.environ/GITHUB_API_KEY
|
| 47 |
+
model: github/Cohere-command-r-plus
|
| 48 |
+
model_name: cohere/command-r-plus
|
| 49 |
+
- litellm_params:
|
| 50 |
+
api_key: os.environ/GITHUB_API_KEY
|
| 51 |
+
model: github/Meta-Llama-3-70B-Instruct
|
| 52 |
+
model_name: meta/llama-3.3-70b
|
| 53 |
+
- litellm_params:
|
| 54 |
+
api_key: os.environ/GITHUB_API_KEY
|
| 55 |
+
model: github/Meta-Llama-3-8B-Instruct
|
| 56 |
+
model_name: meta/llama-3.3-8b
|
generated/groq.yaml
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
- litellm_params:
|
| 2 |
+
api_key: os.environ/GROQ_API_KEY
|
| 3 |
+
model: groq/llama-3.3-70b-versatile
|
| 4 |
+
model_name: meta/llama-3.3-70b
|
| 5 |
+
- litellm_params:
|
| 6 |
+
api_key: os.environ/GROQ_API_KEY
|
| 7 |
+
model: groq/deepseek-r1-distill-llama-70b
|
| 8 |
+
model_name: deepseek/r1-distill-llama-70b
|
generated/huggingface.yaml
ADDED
|
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
- litellm_params:
|
| 2 |
+
model: huggingface/deepseek-ai/DeepSeek-R1-Distill-Qwen-32B
|
| 3 |
+
model_name: deepseek/r1-distill-qwen-32b
|
| 4 |
+
- litellm_params:
|
| 5 |
+
model: huggingface/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
|
| 6 |
+
model_name: deepseek/r1-distill-qwen-1.5b
|
| 7 |
+
- litellm_params:
|
| 8 |
+
model: huggingface/google/gemma-2-2b-it
|
| 9 |
+
model_name: google/gemma-2-2b-it
|
| 10 |
+
- litellm_params:
|
| 11 |
+
model: huggingface/google/gemma-2-9b-it
|
| 12 |
+
model_name: google/gemma-2-9b-it
|
| 13 |
+
- litellm_params:
|
| 14 |
+
model: huggingface/google/gemma-2-27b-it
|
| 15 |
+
model_name: google/gemma-2-27b-it
|
generated/openai.yaml
ADDED
|
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
- litellm_params:
|
| 2 |
+
api_key: os.environ/OPENAI_API_KEY
|
| 3 |
+
model: openai/gpt-4o
|
| 4 |
+
model_name: openai/gpt-4o
|
| 5 |
+
- litellm_params:
|
| 6 |
+
api_key: os.environ/OPENAI_API_KEY
|
| 7 |
+
model: openai/gpt-4o-mini
|
| 8 |
+
model_name: openai/gpt-4o-mini
|
| 9 |
+
- litellm_params:
|
| 10 |
+
api_key: os.environ/OPENAI_API_KEY
|
| 11 |
+
model: openai/o1-mini
|
| 12 |
+
model_name: openai/o1-mini
|
| 13 |
+
- litellm_params:
|
| 14 |
+
api_key: os.environ/OPENAI_API_KEY
|
| 15 |
+
model: openai/o1-preview
|
| 16 |
+
model_name: openai/o1-preview
|
| 17 |
+
- litellm_params:
|
| 18 |
+
api_key: os.environ/OPENAI_API_KEY
|
| 19 |
+
model: openai/gpt-3.5-turbo
|
| 20 |
+
model_name: openai/gpt-3.5-turbo
|
| 21 |
+
- litellm_params:
|
| 22 |
+
api_key: os.environ/OPENAI_API_KEY
|
| 23 |
+
model: openai/text-embedding-3-small
|
| 24 |
+
model_name: openai/text-embedding-3-small
|
| 25 |
+
- litellm_params:
|
| 26 |
+
api_key: os.environ/OPENAI_API_KEY
|
| 27 |
+
model: openai/text-embedding-3-large
|
| 28 |
+
model_name: openai/text-embedding-3-large
|
| 29 |
+
- litellm_params:
|
| 30 |
+
api_key: os.environ/OPENAI_API_KEY
|
| 31 |
+
model: openai/dall-e-3
|
| 32 |
+
model_name: openai/dall-e-3
|
generated/perplexity.yaml
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
- litellm_params:
|
| 2 |
+
api_key: os.environ/PERPLEXITY_API_KEY
|
| 3 |
+
model: perplexity/sonar-pro
|
| 4 |
+
model_name: perplexity/sonar-pro
|
| 5 |
+
- litellm_params:
|
| 6 |
+
api_key: os.environ/PERPLEXITY_API_KEY
|
| 7 |
+
model: perplexity/sonar-reasoning
|
| 8 |
+
model_name: perplexity/sonar-reasoning
|
| 9 |
+
- litellm_params:
|
| 10 |
+
api_key: os.environ/PERPLEXITY_API_KEY
|
| 11 |
+
model: perplexity/sonar
|
| 12 |
+
model_name: perplexity/sonar
|
start.sh
CHANGED
|
@@ -3,42 +3,34 @@ set -euo pipefail
|
|
| 3 |
|
| 4 |
pids=()
|
| 5 |
|
| 6 |
-
|
| 7 |
-
|
| 8 |
-
key_env="$2"
|
| 9 |
|
| 10 |
-
if [[
|
| 11 |
-
>&2 echo "[
|
| 12 |
-
cat
|
| 13 |
-
|
| 14 |
-
litellm_params:
|
| 15 |
-
model: "${model}"
|
| 16 |
-
api_key: "os.environ/${key_env}"
|
| 17 |
-
EOF
|
| 18 |
fi
|
| 19 |
-
}
|
| 20 |
|
| 21 |
-
generate_litellm_provider() {
|
| 22 |
-
provider="$1"
|
| 23 |
key_env="$2"
|
| 24 |
|
| 25 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 26 |
}
|
| 27 |
|
| 28 |
generate_litellm_config() {
|
| 29 |
-
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
|
| 33 |
-
|
| 34 |
-
EOF
|
| 35 |
generate_litellm_provider "openai" "OPENAI_API_KEY"
|
| 36 |
generate_litellm_provider "perplexity" "PERPLEXITY_API_KEY"
|
| 37 |
-
generate_litellm_provider "anthropic" "ANTHROPIC_API_KEY"
|
| 38 |
-
|
| 39 |
-
while read -r model; do
|
| 40 |
-
generate_litellm_model "github/${model}" "GITHUB_API_KEY"
|
| 41 |
-
done </assets/azure-models.txt
|
| 42 |
|
| 43 |
if [[ -n "${LITELLM_MODELS_BASE64:-}" ]]; then
|
| 44 |
echo "${LITELLM_MODELS_BASE64}" | base64 -d
|
|
|
|
| 3 |
|
| 4 |
pids=()
|
| 5 |
|
| 6 |
+
generate_litellm_provider() {
|
| 7 |
+
provider="$1"
|
|
|
|
| 8 |
|
| 9 |
+
if [[ "$#" -lt 2 ]]; then
|
| 10 |
+
>&2 echo "[+] Provider ${provider}: No key needed"
|
| 11 |
+
cat "/assets/${provider}.yaml"
|
| 12 |
+
return
|
|
|
|
|
|
|
|
|
|
|
|
|
| 13 |
fi
|
|
|
|
| 14 |
|
|
|
|
|
|
|
| 15 |
key_env="$2"
|
| 16 |
|
| 17 |
+
if [[ -n "${!key_env:-}" ]]; then
|
| 18 |
+
>&2 echo "[+] Provider ${provider}: Found key ${key_env}"
|
| 19 |
+
cat "/assets/${provider}.yaml"
|
| 20 |
+
return
|
| 21 |
+
fi
|
| 22 |
+
|
| 23 |
+
>&2 echo "[!] Provider ${provider}: No key found (set ${key_env})"
|
| 24 |
}
|
| 25 |
|
| 26 |
generate_litellm_config() {
|
| 27 |
+
echo "model_list:"
|
| 28 |
+
generate_litellm_provider "anthropic" "ANTHROPIC_API_KEY"
|
| 29 |
+
generate_litellm_provider "github" "GITHUB_API_KEY"
|
| 30 |
+
generate_litellm_provider "groq" "GROQ_API_KEY"
|
| 31 |
+
generate_litellm_provider "huggingface"
|
|
|
|
| 32 |
generate_litellm_provider "openai" "OPENAI_API_KEY"
|
| 33 |
generate_litellm_provider "perplexity" "PERPLEXITY_API_KEY"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 34 |
|
| 35 |
if [[ -n "${LITELLM_MODELS_BASE64:-}" ]]; then
|
| 36 |
echo "${LITELLM_MODELS_BASE64}" | base64 -d
|