SarmaHighOnAI commited on
Commit
94491dc
·
verified ·
1 Parent(s): 460fa19

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -8
app.py CHANGED
@@ -1,10 +1,12 @@
1
  import os
 
2
  from fastapi import FastAPI
3
  from pydantic import BaseModel
4
- from huggingface_hub import InferenceClient
5
 
6
  app = FastAPI()
7
- client = InferenceClient(token=os.environ.get("HF_TOKEN"))
 
 
8
 
9
  class Request(BaseModel):
10
  prompt: str
@@ -15,9 +17,7 @@ def home():
15
 
16
  @app.post("/generate")
17
  def generate_text(request: Request):
18
- response = client.chat_completion(
19
- messages=[{"role": "user", "content": request.prompt}],
20
- model="HuggingFaceH4/zephyr-7b-beta",
21
- max_tokens=256
22
- )
23
- return {"response": response.choices[0].message.content}
 
1
  import os
2
+ import requests as req
3
  from fastapi import FastAPI
4
  from pydantic import BaseModel
 
5
 
6
  app = FastAPI()
7
+
8
+ HF_TOKEN = os.environ.get("HF_TOKEN")
9
+ API_URL = "https://api-inference.huggingface.co/models/google/flan-t5-large"
10
 
11
  class Request(BaseModel):
12
  prompt: str
 
17
 
18
  @app.post("/generate")
19
  def generate_text(request: Request):
20
+ headers = {"Authorization": f"Bearer {HF_TOKEN}"}
21
+ payload = {"inputs": request.prompt}
22
+ response = req.post(API_URL, headers=headers, json=payload)
23
+ return {"response": response.json()[0]["generated_text"]}