Watch-Dog-Env / node_modules /@huggingface /tasks /src /inference-providers.ts
Mooizz's picture
Upload folder using huggingface_hub
66980c0 verified
/// This list is for illustration purposes only.
/// in the `tasks` sub-package, we do not need actual strong typing of the inference providers.
const INFERENCE_PROVIDERS = [
"cerebras",
"cohere",
"deepinfra",
"fal-ai",
"fireworks-ai",
"hf-inference",
"hyperbolic",
"ovhcloud",
"replicate",
"sambanova",
"together",
] as const;
export type SnippetInferenceProvider = (typeof INFERENCE_PROVIDERS)[number] | string;
export const HF_HUB_INFERENCE_PROXY_TEMPLATE = `https://router.huggingface.co/{{PROVIDER}}`;
/**
* URL to set as baseUrl in the OpenAI SDK.
*
* TODO(Expose this from InferenceClient in the future?)
*/
export function openAIbaseUrl(provider: SnippetInferenceProvider): string {
const url = HF_HUB_INFERENCE_PROXY_TEMPLATE.replace("{{PROVIDER}}", provider);
return provider === "hf-inference" ? `${url}/v1` : url;
}