Phi2-Fine-Tuning
/
phivenv
/Lib
/site-packages
/huggingface_hub
/inference
/_providers
/featherless_ai.py
| from typing import Any, Dict, Optional, Union | |
| from huggingface_hub.hf_api import InferenceProviderMapping | |
| from huggingface_hub.inference._common import RequestParameters, _as_dict | |
| from ._common import BaseConversationalTask, BaseTextGenerationTask, filter_none | |
| _PROVIDER = "featherless-ai" | |
| _BASE_URL = "https://api.featherless.ai" | |
| class FeatherlessTextGenerationTask(BaseTextGenerationTask): | |
| def __init__(self): | |
| super().__init__(provider=_PROVIDER, base_url=_BASE_URL) | |
| def _prepare_payload_as_dict( | |
| self, inputs: Any, parameters: Dict, provider_mapping_info: InferenceProviderMapping | |
| ) -> Optional[Dict]: | |
| params = filter_none(parameters.copy()) | |
| params["max_tokens"] = params.pop("max_new_tokens", None) | |
| return {"prompt": inputs, **params, "model": provider_mapping_info.provider_id} | |
| def get_response(self, response: Union[bytes, Dict], request_params: Optional[RequestParameters] = None) -> Any: | |
| output = _as_dict(response)["choices"][0] | |
| return { | |
| "generated_text": output["text"], | |
| "details": { | |
| "finish_reason": output.get("finish_reason"), | |
| "seed": output.get("seed"), | |
| }, | |
| } | |
| class FeatherlessConversationalTask(BaseConversationalTask): | |
| def __init__(self): | |
| super().__init__(provider=_PROVIDER, base_url=_BASE_URL) | |