Spaces:
Running
Running
Update app/providers.py
Browse files- app/providers.py +5 -4
app/providers.py
CHANGED
|
@@ -179,16 +179,17 @@ class OpenRouterProvider(BaseProvider):
|
|
| 179 |
|
| 180 |
|
| 181 |
class HuggingFaceProvider(BaseProvider):
|
| 182 |
-
"""HuggingFace Inference API —
|
| 183 |
|
| 184 |
-
|
| 185 |
-
|
|
|
|
| 186 |
"""
|
| 187 |
|
| 188 |
async def complete(self, prompt: str, model: str = None, max_tokens: int = 512) -> str:
|
| 189 |
m = model or self.model
|
| 190 |
data = await self._post(
|
| 191 |
-
f"{self.base_url}/
|
| 192 |
headers={
|
| 193 |
"Authorization": f"Bearer {self.key}",
|
| 194 |
"content-type": "application/json",
|
|
|
|
| 179 |
|
| 180 |
|
| 181 |
class HuggingFaceProvider(BaseProvider):
|
| 182 |
+
"""HuggingFace Inference API — new router endpoint (hf-inference).
|
| 183 |
|
| 184 |
+
Old endpoint /models/{model}/v1/ is deprecated/410 for most models.
|
| 185 |
+
New endpoint: router.huggingface.co/hf-inference/v1/chat/completions
|
| 186 |
+
base_url in .pyfun must be: https://router.huggingface.co/hf-inference/v1
|
| 187 |
"""
|
| 188 |
|
| 189 |
async def complete(self, prompt: str, model: str = None, max_tokens: int = 512) -> str:
|
| 190 |
m = model or self.model
|
| 191 |
data = await self._post(
|
| 192 |
+
f"{self.base_url}/chat/completions", # model im payload, nicht in URL!
|
| 193 |
headers={
|
| 194 |
"Authorization": f"Bearer {self.key}",
|
| 195 |
"content-type": "application/json",
|