Mayur74 commited on
Commit
0c6d327
·
verified ·
1 Parent(s): 1c5ccfd

Create app.py

Browse files
Files changed (1) hide show
  1. app.py +28 -0
app.py ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
3
+
4
+ model_id = "Mayur74/Llama-2-7b-chat-finetune" # your uploaded model
5
+
6
+ # Load model and tokenizer
7
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
8
+ model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", torch_dtype="auto")
9
+
10
+ # Set up pipeline
11
+ pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
12
+
13
+ # Define generation function
14
+ def generate_prompt(prompt):
15
+ output = pipe(prompt, max_new_tokens=300, temperature=0.7)
16
+ return output[0]["generated_text"]
17
+
18
+ # Create Gradio Interface
19
+ demo = gr.Interface(
20
+ fn=generate_prompt,
21
+ inputs=gr.Textbox(lines=5, label="Base Prompt"),
22
+ outputs="text",
23
+ title="🧠 LLaMA 2 Prompt Optimizer",
24
+ description="Enter your prompt and get an optimized version.",
25
+ )
26
+
27
+ # Enable API mode
28
+ demo.launch(share=False, server_name="0.0.0.0", server_port=7860)