madox81 commited on
Commit
6d83ab2
·
verified ·
1 Parent(s): 9d4c13e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +13 -13
app.py CHANGED
@@ -5,12 +5,12 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
5
  from peft import PeftModel
6
 
7
  # --- Configuration ---
8
- BASE_MODEL_ID = "HuggingFaceTB/SmolLM2-1.7B-Instruct"
9
- LORA_REPO_ID = "madox81/SmolLM2-Cyber-Insight_LoRA" # Your LoRA repo
10
 
11
  # --- LLM Class ---
12
  class LLM:
13
- def __init__(self, base_id, lora_id):
14
  print("Loading model...")
15
 
16
  # 1. Device & Dtype
@@ -22,24 +22,24 @@ class LLM:
22
  device_map = "cpu"
23
 
24
  # 2. Load Tokenizer
25
- self.tokenizer = AutoTokenizer.from_pretrained(base_id)
26
  if self.tokenizer.pad_token is None:
27
  self.tokenizer.pad_token = self.tokenizer.eos_token
28
 
29
  # 3. Load Base Model
30
  self.model = AutoModelForCausalLM.from_pretrained(
31
- base_id,
32
  torch_dtype=dtype,
33
  device_map=device_map
34
  )
35
 
36
- # 4. Load LoRA Adapters
37
- print(f"Loading adapters from {lora_id}...")
38
- self.model = PeftModel.from_pretrained(
39
- self.model,
40
- lora_id,
41
- torch_dtype=dtype
42
- )
43
  print("Model loaded!")
44
 
45
  def generate_resp(self, user_input, task_type):
@@ -91,7 +91,7 @@ class LLM:
91
  return response.strip()
92
 
93
  # --- Initialize ---
94
- llm_instance = LLM(BASE_MODEL_ID, LORA_REPO_ID)
95
 
96
  # --- Gradio Interface ---
97
  def process_input(user_input, task_type):
 
5
  from peft import PeftModel
6
 
7
  # --- Configuration ---
8
+ MODEL_ID = "madox81/SmolLM2-Cyber-Insight"
9
+
10
 
11
  # --- LLM Class ---
12
  class LLM:
13
+ def __init__(self, model_id):
14
  print("Loading model...")
15
 
16
  # 1. Device & Dtype
 
22
  device_map = "cpu"
23
 
24
  # 2. Load Tokenizer
25
+ self.tokenizer = AutoTokenizer.from_pretrained(model_id)
26
  if self.tokenizer.pad_token is None:
27
  self.tokenizer.pad_token = self.tokenizer.eos_token
28
 
29
  # 3. Load Base Model
30
  self.model = AutoModelForCausalLM.from_pretrained(
31
+ model_id,
32
  torch_dtype=dtype,
33
  device_map=device_map
34
  )
35
 
36
+ # # 4. Load LoRA Adapters
37
+ # print(f"Loading adapters from {lora_id}...")
38
+ # self.model = PeftModel.from_pretrained(
39
+ # self.model,
40
+ # lora_id,
41
+ # torch_dtype=dtype
42
+ # )
43
  print("Model loaded!")
44
 
45
  def generate_resp(self, user_input, task_type):
 
91
  return response.strip()
92
 
93
  # --- Initialize ---
94
+ llm_instance = LLM(MODEL_ID)
95
 
96
  # --- Gradio Interface ---
97
  def process_input(user_input, task_type):