cpt core 4
Browse files
scripts/cpt_core_model_4.py
CHANGED
|
@@ -41,7 +41,8 @@ model = FastLanguageModel.get_peft_model(
|
|
| 41 |
lora_dropout = 0, # Supports any, but = 0 is optimized
|
| 42 |
bias = "none", # Supports any, but = "none" is optimized
|
| 43 |
# [NEW] "unsloth" uses 30% less VRAM, fits 2x larger batch sizes!
|
| 44 |
-
use_gradient_checkpointing = "unsloth", # True or "unsloth" for very long context
|
|
|
|
| 45 |
random_state = 3407,
|
| 46 |
use_rslora = True, # We support rank stabilized LoRA
|
| 47 |
loftq_config = None, # And LoftQ
|
|
|
|
| 41 |
lora_dropout = 0, # Supports any, but = 0 is optimized
|
| 42 |
bias = "none", # Supports any, but = "none" is optimized
|
| 43 |
# [NEW] "unsloth" uses 30% less VRAM, fits 2x larger batch sizes!
|
| 44 |
+
# use_gradient_checkpointing = "unsloth", # True or "unsloth" for very long context
|
| 45 |
+
use_gradient_checkpointing = False,
|
| 46 |
random_state = 3407,
|
| 47 |
use_rslora = True, # We support rank stabilized LoRA
|
| 48 |
loftq_config = None, # And LoftQ
|