Spaces:
Paused
Paused
Revert "Revert "feat(huggingchat): model config update 25-09-2024 (#1494)""
Browse files- chart/env/prod.yaml +26 -0
chart/env/prod.yaml
CHANGED
|
@@ -134,6 +134,32 @@ envVars:
|
|
| 134 |
}
|
| 135 |
]
|
| 136 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 137 |
{
|
| 138 |
"name": "NousResearch/Hermes-3-Llama-3.1-8B",
|
| 139 |
"description": "Nous Research's latest Hermes 3 release in 8B size. Follows instruction closely.",
|
|
|
|
| 134 |
}
|
| 135 |
]
|
| 136 |
},
|
| 137 |
+
{
|
| 138 |
+
"name": "meta-llama/Llama-3.2-11B-Vision-Instruct",
|
| 139 |
+
"logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/meta-logo.png",
|
| 140 |
+
"description": "The latest multimodal model from Meta! Supports image inputs natively.",
|
| 141 |
+
"websiteUrl": "https://llama.com/",
|
| 142 |
+
"multimodal": true,
|
| 143 |
+
"parameters": {
|
| 144 |
+
"stop": ["<|eot_id|>", "<|im_end|>"],
|
| 145 |
+
"temperature": 0.6,
|
| 146 |
+
"truncate": 14336,
|
| 147 |
+
"max_new_tokens": 1536
|
| 148 |
+
},
|
| 149 |
+
"endpoint": {
|
| 150 |
+
"type": "openai",
|
| 151 |
+
"baseURL": "https://api-inference.huggingface.co/models/meta-llama/Llama-3.2-11B-Vision-Instruct/v1",
|
| 152 |
+
"multimodal": {
|
| 153 |
+
"image": {
|
| 154 |
+
"maxSizeInMB": 10,
|
| 155 |
+
"maxWidth": 560,
|
| 156 |
+
"maxHeight": 560,
|
| 157 |
+
"supportedMimeTypes": ["image/png", "image/jpeg", "image/webp"],
|
| 158 |
+
"preferredMimeType": "image/webp"
|
| 159 |
+
}
|
| 160 |
+
}
|
| 161 |
+
}
|
| 162 |
+
},
|
| 163 |
{
|
| 164 |
"name": "NousResearch/Hermes-3-Llama-3.1-8B",
|
| 165 |
"description": "Nous Research's latest Hermes 3 release in 8B size. Follows instruction closely.",
|