Prajjwalng commited on
Commit
42f5635
·
verified ·
1 Parent(s): aa49720

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -4
app.py CHANGED
@@ -17,12 +17,12 @@ else:
17
  st.warning("HF_TOKEN environment variable not set. Some features may be limited.")
18
 
19
  # Model and Adapter Configuration
20
- model_id = "google/gemma-2b-it" # Base model
21
  adapter_id = "Prajjwalng/gemma_customercare_adapters" #adapter model
22
 
23
  # Initialize model and tokenizer (load only once)
24
  @st.cache_resource
25
- def load_model(model_id, adapter_id):
26
  base_model = AutoModelForCausalLM.from_pretrained(
27
  model_id,
28
  low_cpu_mem_usage=True,
@@ -31,9 +31,8 @@ def load_model(model_id, adapter_id):
31
  device_map={"": 0} if torch.cuda.is_available() else "cpu"
32
  )
33
 
34
- merged_model = PeftModel.from_pretrained(base_model, adapter_id)
35
  tokenizer = AutoTokenizer.from_pretrained(model_id, add_eos_token=True)
36
- return merged_model, tokenizer
37
 
38
  merged_model, tokenizer = load_model(model_id, adapter_id)
39
 
 
17
  st.warning("HF_TOKEN environment variable not set. Some features may be limited.")
18
 
19
  # Model and Adapter Configuration
20
+ model_id = "Prajjwalng/gemma_customer_care" # Base model
21
  adapter_id = "Prajjwalng/gemma_customercare_adapters" #adapter model
22
 
23
  # Initialize model and tokenizer (load only once)
24
  @st.cache_resource
25
+ def load_model(model_id):
26
  base_model = AutoModelForCausalLM.from_pretrained(
27
  model_id,
28
  low_cpu_mem_usage=True,
 
31
  device_map={"": 0} if torch.cuda.is_available() else "cpu"
32
  )
33
 
 
34
  tokenizer = AutoTokenizer.from_pretrained(model_id, add_eos_token=True)
35
+ return base_model
36
 
37
  merged_model, tokenizer = load_model(model_id, adapter_id)
38