aiart-main / app.py
PLAZMAstudios's picture
Upload folder using huggingface_hub
e6905ea verified
import gradio as gr
from transformers import pipeline
# Models
models = {
"text_generator": {"model": "EleutherAI/gpt-neo-1.3B", "task": "text-generation"},
"sentiment": {"model": "cardiffnlp/twitter-roberta-base-sentiment-latest", "task": "sentiment-analysis"},
"dialogue": {"model": "microsoft/DialoGPT-medium", "task": "text-generation"},
"summarizer": {"model": "sshleifer/distilbart-cnn-12-6", "task": "summarization"},
"ner": {"model": "dslim/bert-base-NER", "task": "ner"},
}
# Initialize pipelines
pipelines = {}
for key, info in models.items():
pipelines[key] = pipeline(info["task"], model=info["model"])
def generate_text(prompt, max_length):
result = pipelines["text_generator"](prompt, max_length=int(max_length), num_return_sequences=1)
return result[0]['generated_text']
def analyze_sentiment(text):
result = pipelines["sentiment"](text)
return [{"label": res['label'], "score": f"{res['score']:.2f}"} for res in result]
def converse(message, state):
chat_history_str = state or ""
chat_input_ids = pipelines["dialogue"].tokenizer.encode(chat_history_str + pipelines["dialogue"].tokenizer.eos_token + message, return_tensors='pt')
bot_input_ids = pipelines["dialogue"].model.generate(chat_input_ids, max_length=1000, pad_token_id=pipelines["dialogue"].tokenizer.eos_token_id)
bot_response = pipelines["dialogue"].tokenizer.decode(bot_input_ids[:, chat_input_ids.shape[-1]:][0], skip_special_tokens=True)
chat_history = chat_history_str + pipelines["dialogue"].tokenizer.eos_token + message + pipelines["dialogue"].tokenizer.eos_token + bot_response
return bot_response, chat_history
def summarize_text(text, max_length, min_length):
result = pipelines["summarizer"](text, max_length=int(max_length), min_length=int(min_length), do_sample=False)
return result[0]['summary_text']
def ner_analysis(text):
result = pipelines["ner"](text, aggregation_strategy="max")
formatted = [f"{res['entity_group']}: {res['word']} ({res['score']:.2f})" for res in result]
return "\n".join(formatted)
with gr.Blocks() as demo:
gr.Markdown("# AI Art Main Interface")
gr.Markdown("Use various AI models below")
with gr.Tab("Text Generator"):
gr.Markdown("## Text Generator (EleutherAI/gpt-neo-1.3B)")
gen_prompt = gr.Textbox(label="Prompt")
gen_len = gr.Slider(10, 200, 50, label="Max Length")
gen_btn = gr.Button()
gen_out = gr.Textbox(label="Output")
gen_btn.click(generate_text, [gen_prompt, gen_len], gen_out)
with gr.Tab("Sentiment Analysis"):
gr.Markdown("## Sentiment Analyzer (cardiffnlp/twitter-roberta-base-sentiment-latest)")
sent_text = gr.Textbox(label="Text")
sent_btn = gr.Button()
sent_out = gr.JSON(label="Result")
sent_btn.click(analyze_sentiment, sent_text, sent_out)
with gr.Tab("Dialogue Bot"):
gr.Markdown("## Dialogue Bot (microsoft/DialoGPT-medium)")
chat_state = gr.State()
chat_msg = gr.Textbox(label="Message")
chat_send = gr.Button()
chat_out = gr.Chatbot()
chat_send.click(converse, [chat_msg, chat_state], [chat_out, chat_state])
with gr.Tab("Text Summarizer"):
gr.Markdown("## Text Summarizer (sshleifer/distilbart-cnn-12-6)")
sum_text = gr.Textbox(lines=5, label="Text")
sum_max = gr.Slider(10, 200, 100, label="Max Length")
sum_min = gr.Slider(5, 50, 10, label="Min Length")
sum_btn = gr.Button()
sum_out = gr.Textbox(label="Summary")
sum_btn.click(summarize_text, [sum_text, sum_max, sum_min], sum_out)
with gr.Tab("Named Entity Recognition"):
gr.Markdown("## Named Entity Recognition (dslim/bert-base-NER)")
ner_text = gr.Textbox(lines=3, label="Text")
ner_btn = gr.Button()
ner_out = gr.Textbox(lines=5, label="Entities")
ner_btn.click(ner_analysis, ner_text, ner_out)
if __name__ == "__main__":
demo.launch()