|
|
import gradio as gr |
|
|
from transformers import ( |
|
|
GPT2LMHeadModel, GPT2Tokenizer, |
|
|
AutoModelForCausalLM, AutoTokenizer |
|
|
) |
|
|
|
|
|
|
|
|
gpt2_model = GPT2LMHeadModel.from_pretrained("gpt2") |
|
|
gpt2_tokenizer = GPT2Tokenizer.from_pretrained("gpt2") |
|
|
|
|
|
|
|
|
bloom_model = AutoModelForCausalLM.from_pretrained("bigscience/bloom-560m") |
|
|
bloom_tokenizer = AutoTokenizer.from_pretrained("bigscience/bloom-560m") |
|
|
|
|
|
|
|
|
def generate_text(prompt, model_name): |
|
|
if model_name == "🧠 GPT2": |
|
|
inputs = gpt2_tokenizer.encode(prompt, return_tensors="pt") |
|
|
output = gpt2_model.generate(inputs, max_length=100) |
|
|
return gpt2_tokenizer.decode(output[0], skip_special_tokens=True) |
|
|
|
|
|
elif model_name == "🌸 Bloom-560M": |
|
|
inputs = bloom_tokenizer(prompt, return_tensors="pt") |
|
|
output = bloom_model.generate(inputs["input_ids"], max_length=100) |
|
|
return bloom_tokenizer.decode(output[0], skip_special_tokens=True) |
|
|
|
|
|
|
|
|
with gr.Blocks(css=""" |
|
|
body { background-color: #FFFACD; } |
|
|
h1 { color: brown !important; } |
|
|
""") as demo: |
|
|
|
|
|
gr.Markdown("<h1 style='text-align: center;'>LLM for Content Generation</h1>") |
|
|
gr.Markdown("<div style='text-align: center;'>Generate high-quality text using powerful LLMs</div>") |
|
|
|
|
|
with gr.Row(): |
|
|
with gr.Column(): |
|
|
prompt = gr.Textbox(label="Enter a topic or prompt") |
|
|
model_choice = gr.Radio(["🧠 GPT2", "🌸 Bloom-560M"], label="Choose a Model") |
|
|
submit = gr.Button("Generate") |
|
|
|
|
|
with gr.Column(): |
|
|
output = gr.Textbox(label="Generated Text", lines=10) |
|
|
|
|
|
submit.click(fn=generate_text, inputs=[prompt, model_choice], outputs=output) |
|
|
|
|
|
gr.Markdown("<div style='text-align: center; color: brown; margin-top: 30px;'>Designed by Mehak Mazhar</div>") |
|
|
|
|
|
demo.launch() |
|
|
|