kouki321 commited on
Commit
5d96d6a
·
verified ·
1 Parent(s): 9cdd571

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -9
app.py CHANGED
@@ -1,23 +1,22 @@
1
- from fastapi import FastAPI, Request
2
  #from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
3
- import torch
4
  from transformers import AutoTokenizer, AutoModelForSeq2SeqLM, pipeline
5
 
6
  app = FastAPI()
7
-
8
  model_id = "google/flan-t5-small"
 
9
  #"unsloth/mistral-7b-v0.2-bnb-4bit"
10
  #deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
11
- tokenizer = AutoTokenizer.from_pretrained(model_id,trust_remote_code=True)
12
- model = AutoModelForSeq2SeqLM.from_pretrained(model_id
13
- , trust_remote_code=True
14
- )
15
 
16
- generator = pipeline("text-generation", model=model, tokenizer=tokenizer)
 
 
 
17
 
18
  @app.post("/generate")
19
  async def generate(request: Request):
20
  data = await request.json()
21
  prompt = data.get("prompt", "")
22
- result = generator(prompt, max_new_tokens=100, do_sample=True)[0]["generated_text"]
23
  return {"output": result}
 
 
 
1
  #from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
2
+ from fastapi import FastAPI, Request
3
  from transformers import AutoTokenizer, AutoModelForSeq2SeqLM, pipeline
4
 
5
  app = FastAPI()
 
6
  model_id = "google/flan-t5-small"
7
+
8
  #"unsloth/mistral-7b-v0.2-bnb-4bit"
9
  #deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
 
 
 
 
10
 
11
+
12
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
13
+ model = AutoModelForSeq2SeqLM.from_pretrained(model_id)
14
+ generator = pipeline("text2text-generation", model=model, tokenizer=tokenizer)
15
 
16
  @app.post("/generate")
17
  async def generate(request: Request):
18
  data = await request.json()
19
  prompt = data.get("prompt", "")
20
+ result = generator(prompt, max_new_tokens=100)[0]["generated_text"]
21
  return {"output": result}
22
+