Update app.py
Browse files
app.py
CHANGED
|
@@ -18,7 +18,7 @@ print(f"stage ${current_num}")
|
|
| 18 |
|
| 19 |
api = HfApi(token=hf_token)
|
| 20 |
# models = f"dad1909/cybersentinal-2.0-{current_num}"
|
| 21 |
-
model_base = "unsloth/
|
| 22 |
|
| 23 |
print("Starting model and tokenizer loading...")
|
| 24 |
|
|
@@ -28,8 +28,7 @@ model, tokenizer = FastLanguageModel.from_pretrained(
|
|
| 28 |
max_seq_length=max_seq_length,
|
| 29 |
dtype=dtype,
|
| 30 |
load_in_4bit=load_in_4bit,
|
| 31 |
-
token=hf_token
|
| 32 |
-
load_in_8bit_fp32_cpu_offload=True, # Enable CPU offloading
|
| 33 |
)
|
| 34 |
|
| 35 |
print("Model and tokenizer loaded successfully.")
|
|
|
|
| 18 |
|
| 19 |
api = HfApi(token=hf_token)
|
| 20 |
# models = f"dad1909/cybersentinal-2.0-{current_num}"
|
| 21 |
+
model_base = "unsloth/gemma-2-27b-bnb-4bit"
|
| 22 |
|
| 23 |
print("Starting model and tokenizer loading...")
|
| 24 |
|
|
|
|
| 28 |
max_seq_length=max_seq_length,
|
| 29 |
dtype=dtype,
|
| 30 |
load_in_4bit=load_in_4bit,
|
| 31 |
+
token=hf_token
|
|
|
|
| 32 |
)
|
| 33 |
|
| 34 |
print("Model and tokenizer loaded successfully.")
|