condense.ai / app.py
aliMohammad16's picture
Update app.py
366c751 verified
import gradio as gr
from transformers import AutoTokenizer, AutoModelForCausalLM
# Load DeepSeek model (replace with your actual model name)
MODEL_NAME = "deepseek-ai/deepseek-coder-6.7b-instruct"
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
model = AutoModelForCausalLM.from_pretrained(MODEL_NAME)
def summarize(text):
# Customize your summarization prompt
prompt = f"Summarize this text concisely:\n\n{text}\n\nSummary:"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_length=500)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Create Gradio interface
with gr.Blocks() as demo:
gr.Markdown("# Text Summarizer")
with gr.Row():
with gr.Column():
input_text = gr.Textbox(label="Input Text", lines=10)
submit_btn = gr.Button("Summarize")
with gr.Column():
output_text = gr.Textbox(label="Summary", lines=10)
submit_btn.click(fn=summarize, inputs=input_text, outputs=output_text)
# Launch with API mode enabled
demo.launch(api_mode=True, server_name="0.0.0.0", server_port=7860)