Upload scripts/train_alizee_v2_stage1_sft.py with huggingface_hub
Browse files
scripts/train_alizee_v2_stage1_sft.py
CHANGED
|
@@ -71,10 +71,12 @@ print("\n📝 Loading tokenizer...")
|
|
| 71 |
tokenizer = AutoTokenizer.from_pretrained(
|
| 72 |
BASE_MODEL,
|
| 73 |
trust_remote_code=True,
|
|
|
|
| 74 |
)
|
| 75 |
if tokenizer.pad_token is None:
|
| 76 |
tokenizer.pad_token = tokenizer.eos_token
|
| 77 |
tokenizer.padding_side = "right"
|
|
|
|
| 78 |
|
| 79 |
# QLoRA quantization config
|
| 80 |
print("\n⚙️ Configuring 4-bit quantization...")
|
|
|
|
| 71 |
tokenizer = AutoTokenizer.from_pretrained(
|
| 72 |
BASE_MODEL,
|
| 73 |
trust_remote_code=True,
|
| 74 |
+
model_max_length=MAX_SEQ_LENGTH, # Set max length to prevent overflow warnings
|
| 75 |
)
|
| 76 |
if tokenizer.pad_token is None:
|
| 77 |
tokenizer.pad_token = tokenizer.eos_token
|
| 78 |
tokenizer.padding_side = "right"
|
| 79 |
+
tokenizer.truncation_side = "right" # Truncate from the right for long sequences
|
| 80 |
|
| 81 |
# QLoRA quantization config
|
| 82 |
print("\n⚙️ Configuring 4-bit quantization...")
|