Update app.py
Browse files
app.py
CHANGED
|
@@ -6,10 +6,6 @@ import sys
|
|
| 6 |
from dotenv import load_dotenv, dotenv_values
|
| 7 |
load_dotenv()
|
| 8 |
|
| 9 |
-
|
| 10 |
-
|
| 11 |
-
|
| 12 |
-
|
| 13 |
# initialize the client
|
| 14 |
client = OpenAI(
|
| 15 |
base_url="https://api-inference.huggingface.co/v1",
|
|
@@ -18,29 +14,17 @@ client = OpenAI(
|
|
| 18 |
|
| 19 |
# Create supported models
|
| 20 |
model_links = {
|
| 21 |
-
"Meta-Llama-2-7B-hf":"meta-llama/Llama-2-7b-hf",
|
| 22 |
-
"Google-gemma-2-9b":"google/gemma-2-9b",
|
| 23 |
-
"Meta-Llama-3.1-70B-Instruct": "meta-llama/Meta-Llama-3.1-70B-Instruct",
|
| 24 |
-
"Meta-Llama-3.1-8B-Instruct": "meta-llama/Meta-Llama-3.1-8B-Instruct",
|
| 25 |
-
"Meta-Llama-3.1-405B-Instruct-FP8": "meta-llama/Meta-Llama-3.1-405B-Instruct-FP8",
|
| 26 |
-
"Meta-Llama-3.1-405B-Instruct": "meta-llama/Meta-Llama-3.1-405B-Instruct",
|
| 27 |
"Mistral-Nemo-Instruct-2410": "mistralai/Ministral-8B-Instruct-2410",
|
| 28 |
"Text-to-IMG-FLUX.1-dev": "black-forest-labs/FLUX.1-dev",
|
| 29 |
"Text-to-IMG-NSFW-gen-v2": "UnfilteredAI/NSFW-gen-v2",
|
| 30 |
"C4ai-command-r-plus": "CohereForAI/c4ai-command-r-plus",
|
| 31 |
"Aya-23-35B": "CohereForAI/aya-23-35B",
|
| 32 |
-
"Zephyr-orpo-141b-A35b-v0.1": "HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1",
|
| 33 |
"Mixtral-8x7B-Instruct-v0.1": "mistralai/Mixtral-8x7B-Instruct-v0.1",
|
| 34 |
-
"Codestral-22B-v0.1": "mistralai/Codestral-22B-v0.1",
|
| 35 |
"Nous-Hermes-2-Mixtral-8x7B-DPO": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
|
| 36 |
-
"Yi-1.5-34B-Chat": "01-ai/Yi-1.5-34B-Chat",
|
| 37 |
-
"Gemma-2-27b-it": "google/gemma-2-27b-it",
|
| 38 |
-
"Meta-Llama-2-70B-Chat-HF": "meta-llama/Llama-2-70b-chat-hf",
|
| 39 |
-
"Text-to-IMG-ByteDance/SDXL-Lightning": "ByteDance/SDXL-Lightning",
|
| 40 |
-
"Meta-Llama-2-13B-Chat-HF": "meta-llama/Llama-2-13b-chat-hf",
|
| 41 |
"Mistral-7B-Instruct-v0.1": "mistralai/Mistral-7B-Instruct-v0.1",
|
| 42 |
"Mistral-7B-Instruct-v0.2": "mistralai/Mistral-7B-Instruct-v0.2",
|
| 43 |
"Mistral-7B-Instruct-v0.3": "mistralai/Mistral-7B-Instruct-v0.3",
|
|
|
|
| 44 |
"Falcon-7b-Instruct": "tiiuae/falcon-7b-instruct",
|
| 45 |
"Starchat2-15b-v0.1": "HuggingFaceH4/starchat2-15b-v0.1",
|
| 46 |
"Gemma-1.1-7b-it": "google/gemma-1.1-7b-it",
|
|
|
|
| 6 |
from dotenv import load_dotenv, dotenv_values
|
| 7 |
load_dotenv()
|
| 8 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
# initialize the client
|
| 10 |
client = OpenAI(
|
| 11 |
base_url="https://api-inference.huggingface.co/v1",
|
|
|
|
| 14 |
|
| 15 |
# Create supported models
|
| 16 |
model_links = {
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
"Mistral-Nemo-Instruct-2410": "mistralai/Ministral-8B-Instruct-2410",
|
| 18 |
"Text-to-IMG-FLUX.1-dev": "black-forest-labs/FLUX.1-dev",
|
| 19 |
"Text-to-IMG-NSFW-gen-v2": "UnfilteredAI/NSFW-gen-v2",
|
| 20 |
"C4ai-command-r-plus": "CohereForAI/c4ai-command-r-plus",
|
| 21 |
"Aya-23-35B": "CohereForAI/aya-23-35B",
|
|
|
|
| 22 |
"Mixtral-8x7B-Instruct-v0.1": "mistralai/Mixtral-8x7B-Instruct-v0.1",
|
|
|
|
| 23 |
"Nous-Hermes-2-Mixtral-8x7B-DPO": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 24 |
"Mistral-7B-Instruct-v0.1": "mistralai/Mistral-7B-Instruct-v0.1",
|
| 25 |
"Mistral-7B-Instruct-v0.2": "mistralai/Mistral-7B-Instruct-v0.2",
|
| 26 |
"Mistral-7B-Instruct-v0.3": "mistralai/Mistral-7B-Instruct-v0.3",
|
| 27 |
+
"Mistral-Small-Instruct-2409": "mistralai/Mistral-Small-Instruct-2409",
|
| 28 |
"Falcon-7b-Instruct": "tiiuae/falcon-7b-instruct",
|
| 29 |
"Starchat2-15b-v0.1": "HuggingFaceH4/starchat2-15b-v0.1",
|
| 30 |
"Gemma-1.1-7b-it": "google/gemma-1.1-7b-it",
|