gk2410 commited on
Commit
7804499
·
verified ·
1 Parent(s): ab1e95e

Update llm.py

Browse files
Files changed (1) hide show
  1. llm.py +23 -14
llm.py CHANGED
@@ -1,20 +1,29 @@
1
- import requests
2
  import os
 
 
 
 
3
 
4
- API_URL = "https://api-inference.huggingface.co/models/mistralai/Mistral-7B-Instruct-v0.2"
5
- HEADERS = {
6
- "Authorization": f"Bearer {os.environ['HF_TOKEN']}"
7
- }
 
8
 
9
- def ask_llm(prompt):
10
  payload = {
11
  "inputs": prompt,
12
- "parameters": {
13
- "max_new_tokens": 300,
14
- "temperature": 0.4,
15
- "return_full_text": False
16
- }
17
  }
18
- r = requests.post(API_URL, headers=HEADERS, json=payload, timeout=120)
19
- r.raise_for_status()
20
- return r.json()[0]["generated_text"]
 
 
 
 
 
 
 
 
 
 
 
1
  import os
2
+ import requests
3
+
4
+ HF_MODEL_URL = "https://api-inference.huggingface.co/models/mistralai/Mistral-7B-Instruct-v0.2"
5
+ HF_TOKEN = os.environ.get("HF_TOKEN")
6
 
7
+ HEADERS = {"Authorization": f"Bearer {HF_TOKEN}"}
8
+
9
+ def ask_llm(prompt: str, max_tokens: int = 300) -> str:
10
+ if not HF_TOKEN:
11
+ return "Error: HF_TOKEN environment variable not set."
12
 
 
13
  payload = {
14
  "inputs": prompt,
15
+ "parameters": {"max_new_tokens": max_tokens},
16
+ "options": {"wait_for_model": True}
 
 
 
17
  }
18
+
19
+ try:
20
+ response = requests.post(HF_MODEL_URL, headers=HEADERS, json=payload, timeout=120)
21
+ response.raise_for_status()
22
+ result = response.json()
23
+
24
+ if isinstance(result, list) and "generated_text" in result[0]:
25
+ return result[0]["generated_text"]
26
+ else:
27
+ return str(result)
28
+ except requests.exceptions.RequestException as e:
29
+ return f"Error querying LLM: {e}"