Spaces:
Sleeping
Sleeping
| import os | |
| import requests | |
| from llm_params import LlmParams | |
| class Llm: | |
| params: LlmParams = None | |
| def __init__(self, params): | |
| self.llm_api_endpoint = os.environ.get("LLM_API_ENDPOINT", "") | |
| self.params = params | |
| def predict(self, prompt: str = None): | |
| response = requests.post( | |
| url=self.params.url+'/completion', | |
| json={'prompt': ' [INST] ' + prompt + ' [/INST]', | |
| 'temperature': 0.0, | |
| 'n_predict': 2500.0, | |
| 'top_p': 0.95, | |
| 'min_p': 0.05, | |
| 'repeat_penalty': 1.2, | |
| 'stop': []}) | |
| answer = response.json()['content'] | |
| return answer | |
| def query_llamacpp(self, prompt: str): | |
| response = requests.post( | |
| url=self.params.url+'/completion', | |
| json={'prompt': ' [INST] ' + prompt + ' [/INST]', | |
| 'temperature': 0.0, | |
| 'n_predict': 2500.0, | |
| 'top_p': 0.95, | |
| 'min_p': 0.05, | |
| 'repeat_penalty': 1.2, | |
| 'stop': []}) | |
| answer = response.json()['content'] | |
| return answer |