Monimoy commited on
Commit
5571dd0
·
verified ·
1 Parent(s): fb5822c

Delete app.py

Browse files
Files changed (1) hide show
  1. app.py +0 -55
app.py DELETED
@@ -1,55 +0,0 @@
1
- import torch
2
- from transformers import AutoModelForCausalLM, AutoTokenizer
3
- import openvino_genai
4
- import gradio as gr
5
-
6
-
7
- # Base Phi-2 model name
8
- #base_model_name = "microsoft/phi-2"
9
- base_model_name = "Monimoy/openvino_phi2"
10
- peft_model_path = "./phi2-openassistant-lora-final"
11
-
12
-
13
-
14
- device = 'CPU' # GPU can be used as well
15
- adapter = openvino_genai.Adapter(peft_model_path)
16
- adapter_config = openvino_genai.AdapterConfig(adapter)
17
- pipe = openvino_genai.LLMPipeline(model=base_model_name, device=device, adapters=adapter_config) # register all required adapters here
18
-
19
- print("Generate with LoRA adapter and alpha set to 0.75:")
20
- #print(pipe.generate(args.prompt, max_new_tokens=100, adapters=openvino_genai.AdapterConfig(adapter, 0.75)))
21
-
22
- # Load tokenizer
23
- tokenizer = AutoTokenizer.from_pretrained(base_model_name)
24
- tokenizer.pad_token = tokenizer.eos_token
25
-
26
- # Define prediction function
27
- def generate_response(prompt):
28
- #inputs = tokenizer(prompt, return_tensors="pt").to(device)
29
- #with torch.no_grad():
30
- # output = model.generate(**inputs, max_new_tokens=100, do_sample=True, temperature=0.7)
31
- #return tokenizer.decode(output[0], skip_special_tokens=True)
32
- return pipe.generate(prompt, max_new_tokens=100, adapters=openvino_genai.AdapterConfig(adapter, 0.75))
33
-
34
-
35
- # Define example prompts
36
- examples = [
37
- ["What is machine learning?"],
38
- ["Explain quantum mechanics in simple terms."],
39
- ["Write a short story about a robot discovering emotions."],
40
- ["Summarize the theory of relativity."]
41
- ]
42
-
43
- # Create Gradio UI
44
- iface = gr.Interface(
45
- fn=generate_response,
46
- inputs=gr.Textbox(lines=2, placeholder="Enter your prompt here..."),
47
- outputs=gr.Textbox(),
48
- title="Phi-2 LoRA Model",
49
- description="A fine-tuned Phi-2 model with LoRA running on Hugging Face Spaces (CPU optimized).",
50
- examples=examples,
51
- )
52
-
53
- # Launch Gradio app
54
- iface.launch()
55
-