Monimoy commited on
Commit
7ea9b97
·
verified ·
1 Parent(s): 7cd74fe

Upload app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -15,7 +15,7 @@ peft_model_path = "./phi2-openassistant-lora-final"
15
  # Load the base model with 4-bit quantization
16
  #bnb_config = BitsAndBytesConfig(load_in_4bit=True) # Ensure compatibility
17
  #base_model = AutoModelForCausalLM.from_pretrained(base_model_name, quantization_config=bnb_config, device_map={"": device})
18
- #base_model = AutoModelForCausalLM.from_pretrained(base_model_name, torch_dtype=torch.float32, device_map={"": device})
19
  #base_model = AutoModelForCausalLM.from_pretrained(base_model_name, load_in_4bit=True, device_map={"": device})
20
  #bnb_config = BitsAndBytesConfig(
21
  #load_in_4bit=True,
@@ -25,7 +25,7 @@ peft_model_path = "./phi2-openassistant-lora-final"
25
  #)
26
  #base_model = AutoModelForCausalLM.from_pretrained(base_model_name, quantization_config=bnb_config, device_map={"": device})
27
  #base_model = AutoModelForCausalLM.from_pretrained(base_model_name, model_type="llama", device_map={"": device}, quantization="4bit")
28
- base_model = AutoModelForCausalLM.from_pretrained(base_model_name, load_in_4bit=True, device_map={"": device}, torch_dtype=torch.float16)
29
 
30
 
31
 
 
15
  # Load the base model with 4-bit quantization
16
  #bnb_config = BitsAndBytesConfig(load_in_4bit=True) # Ensure compatibility
17
  #base_model = AutoModelForCausalLM.from_pretrained(base_model_name, quantization_config=bnb_config, device_map={"": device})
18
+ base_model = AutoModelForCausalLM.from_pretrained(base_model_name, torch_dtype=torch.float32, device_map={"": device})
19
  #base_model = AutoModelForCausalLM.from_pretrained(base_model_name, load_in_4bit=True, device_map={"": device})
20
  #bnb_config = BitsAndBytesConfig(
21
  #load_in_4bit=True,
 
25
  #)
26
  #base_model = AutoModelForCausalLM.from_pretrained(base_model_name, quantization_config=bnb_config, device_map={"": device})
27
  #base_model = AutoModelForCausalLM.from_pretrained(base_model_name, model_type="llama", device_map={"": device}, quantization="4bit")
28
+ #base_model = AutoModelForCausalLM.from_pretrained(base_model_name, load_in_4bit=True, device_map={"": device}, torch_dtype=torch.float16)
29
 
30
 
31