File size: 814 Bytes
b44cc33
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e246d3c
b44cc33
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
"""

LLM + Hugging face Model + Deployment using Docker + deployment on Hugging face spaces



"""


from fastapi import FastAPI
from transformers import pipeline
import os

##create a new fastapi app
app=FastAPI()



#Initialize the text generation pipeline
# loading model from the hugging face model hub

pipe = pipeline("text2text-generation", model="google/flan-t5-small")


@app.get("/")
def home():
    return {"message":"Welcome to the Text Generation API"}



@app.get("/generate")
# '''text --> prompt'''

def generate(text:str):
    output = pipe(text, do_sample=True, temperature=0.7)

    #retuen the fgenerated text in json response
    return{"output":output[0]['generate']}