Springboardmen commited on
Commit
b040724
·
verified ·
1 Parent(s): 8416321

Update model_api.py

Browse files
Files changed (1) hide show
  1. model_api.py +17 -29
model_api.py CHANGED
@@ -1,37 +1,25 @@
1
- import requests
2
  import os
3
- import time
4
-
5
- API_URL = "https://router.huggingface.co/hf-inference/models/mistralai/Mistral-7B-Instruct-v0.2"
6
-
7
- HF_TOKEN = os.getenv("HF_TOKEN")
8
-
9
- headers = {
10
- "Authorization": f"Bearer {HF_TOKEN}"
11
- }
12
 
13
  def query_model(prompt):
14
- payload = {
15
- "inputs": prompt,
16
- "parameters": {
17
- "max_new_tokens": 500,
18
- "temperature": 0.7
19
- }
20
- }
21
-
22
- response = requests.post(API_URL, headers=headers, json=payload)
23
 
24
- # Check status code first
25
- if response.status_code != 200:
26
- return f"HTTP Error {response.status_code}: {response.text}"
27
 
 
 
 
 
28
 
29
- try:
30
- result = response.json()
31
- except Exception:
32
- return f"Invalid response received: {response.text}"
 
33
 
34
- if isinstance(result, list):
35
- return result[0]["generated_text"]
36
 
37
- return f"Unexpected response: {result}"
 
 
1
+ from huggingface_hub import InferenceClient
2
  import os
 
 
 
 
 
 
 
 
 
3
 
4
  def query_model(prompt):
5
+ try:
6
+ HF_TOKEN = os.getenv("HF_TOKEN")
 
 
 
 
 
 
 
7
 
8
+ if not HF_TOKEN:
9
+ return "⚠ HF_TOKEN not found."
 
10
 
11
+ client = InferenceClient(
12
+ model="mistralai/Mistral-7B-Instruct-v0.2",
13
+ token=HF_TOKEN
14
+ )
15
 
16
+ response = client.text_generation(
17
+ prompt,
18
+ max_new_tokens=300,
19
+ temperature=0.7
20
+ )
21
 
22
+ return response
 
23
 
24
+ except Exception as e:
25
+ return f"Error: {str(e)}"