SarmaHighOnAI commited on
Commit
f158928
·
verified ·
1 Parent(s): e5b92e5

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -7
app.py CHANGED
@@ -4,10 +4,7 @@ from pydantic import BaseModel
4
  from huggingface_hub import InferenceClient
5
 
6
  app = FastAPI()
7
- client = InferenceClient(
8
- provider="hf-inference",
9
- api_key=os.environ.get("HF_TOKEN")
10
- )
11
 
12
  class Request(BaseModel):
13
  prompt: str
@@ -19,9 +16,9 @@ def home():
19
  @app.post("/generate")
20
  def generate_text(request: Request):
21
  messages = [{"role": "user", "content": request.prompt}]
22
- completion = client.chat.completions.create(
23
- model="Qwen/Qwen2.5-72B-Instruct",
24
  messages=messages,
 
25
  max_tokens=500
26
  )
27
- return {"response": completion.choices[0].message.content}
 
4
  from huggingface_hub import InferenceClient
5
 
6
  app = FastAPI()
7
+ client = InferenceClient(api_key=os.environ.get("HF_TOKEN"))
 
 
 
8
 
9
  class Request(BaseModel):
10
  prompt: str
 
16
  @app.post("/generate")
17
  def generate_text(request: Request):
18
  messages = [{"role": "user", "content": request.prompt}]
19
+ response = client.chat_completion(
 
20
  messages=messages,
21
+ model="HuggingFaceTB/SmolLM2-1.7B-Instruct",
22
  max_tokens=500
23
  )
24
+ return {"response": response.choices[0].message.content}