|
|
import type { WidgetType } from "@huggingface/tasks"; |
|
|
import type { InferenceProvider, InferenceProviderOrPolicy, ModelId } from "../types.js"; |
|
|
export declare const inferenceProviderMappingCache: Map<string, Partial<Record<"black-forest-labs" | "cerebras" | "cohere" | "fal-ai" | "featherless-ai" | "fireworks-ai" | "groq" | "hf-inference" | "hyperbolic" | "nebius" | "novita" | "nscale" | "openai" | "ovhcloud" | "replicate" | "sambanova" | "together", Omit<InferenceProviderModelMapping, "hfModelId">>>>; |
|
|
export type InferenceProviderMapping = Partial<Record<InferenceProvider, Omit<InferenceProviderModelMapping, "hfModelId">>>; |
|
|
export interface InferenceProviderModelMapping { |
|
|
adapter?: string; |
|
|
adapterWeightsPath?: string; |
|
|
hfModelId: ModelId; |
|
|
providerId: string; |
|
|
status: "live" | "staging"; |
|
|
task: WidgetType; |
|
|
} |
|
|
export declare function fetchInferenceProviderMappingForModel(modelId: ModelId, accessToken?: string, options?: { |
|
|
fetch?: (input: RequestInfo, init?: RequestInit) => Promise<Response>; |
|
|
}): Promise<InferenceProviderMapping>; |
|
|
export declare function getInferenceProviderMapping(params: { |
|
|
accessToken?: string; |
|
|
modelId: ModelId; |
|
|
provider: InferenceProvider; |
|
|
task: WidgetType; |
|
|
}, options: { |
|
|
fetch?: (input: RequestInfo, init?: RequestInit) => Promise<Response>; |
|
|
}): Promise<InferenceProviderModelMapping | null>; |
|
|
export declare function resolveProvider(provider?: InferenceProviderOrPolicy, modelId?: string, endpointUrl?: string): Promise<InferenceProvider>; |
|
|
|