SarmaHighOnAI commited on
Commit
7cf4b3d
·
verified ·
1 Parent(s): 6fc59ff

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +12 -16
app.py CHANGED
@@ -1,12 +1,13 @@
1
  import os
2
- import requests as req
3
  from fastapi import FastAPI
4
  from pydantic import BaseModel
 
5
 
6
  app = FastAPI()
7
-
8
- HF_TOKEN = os.environ.get("HF_TOKEN")
9
- API_URL = "https://api-inference.huggingface.co/models/google/flan-t5-large"
 
10
 
11
  class Request(BaseModel):
12
  prompt: str
@@ -17,15 +18,10 @@ def home():
17
 
18
  @app.post("/generate")
19
  def generate_text(request: Request):
20
- headers = {"Authorization": f"Bearer {HF_TOKEN}"}
21
- payload = {"inputs": request.prompt}
22
- response = req.post(API_URL, headers=headers, json=payload)
23
- result = response.json()
24
-
25
- # Handle different response formats
26
- if isinstance(result, list) and len(result) > 0:
27
- return {"response": result[0].get("generated_text", str(result))}
28
- elif isinstance(result, dict):
29
- return {"response": result.get("generated_text", str(result))}
30
- else:
31
- return {"response": str(result)}
 
1
  import os
 
2
  from fastapi import FastAPI
3
  from pydantic import BaseModel
4
+ from huggingface_hub import InferenceClient
5
 
6
  app = FastAPI()
7
+ client = InferenceClient(
8
+ provider="hf-inference",
9
+ api_key=os.environ.get("HF_TOKEN")
10
+ )
11
 
12
  class Request(BaseModel):
13
  prompt: str
 
18
 
19
  @app.post("/generate")
20
  def generate_text(request: Request):
21
+ messages = [{"role": "user", "content": request.prompt}]
22
+ completion = client.chat.completions.create(
23
+ model="Qwen/Qwen2.5-72B-Instruct",
24
+ messages=messages,
25
+ max_tokens=500
26
+ )
27
+ return {"response": completion.choices[0].message.content}