Upload train_question_generator.py with huggingface_hub
Browse files
train_question_generator.py
CHANGED
|
@@ -57,7 +57,7 @@ peft_config = LoraConfig(
|
|
| 57 |
task_type="CAUSAL_LM",
|
| 58 |
)
|
| 59 |
|
| 60 |
-
# Training config -
|
| 61 |
training_args = SFTConfig(
|
| 62 |
output_dir="./question-generator-output",
|
| 63 |
num_train_epochs=2,
|
|
@@ -73,8 +73,6 @@ training_args = SFTConfig(
|
|
| 73 |
push_to_hub=True,
|
| 74 |
hub_model_id="KevinKeller/cognitive-question-generator-qwen2.5-7b",
|
| 75 |
report_to="none",
|
| 76 |
-
max_seq_length=8192,
|
| 77 |
-
gradient_checkpointing=True,
|
| 78 |
)
|
| 79 |
|
| 80 |
print("Starting training...")
|
|
@@ -85,6 +83,7 @@ trainer = SFTTrainer(
|
|
| 85 |
peft_config=peft_config,
|
| 86 |
tokenizer=tokenizer,
|
| 87 |
args=training_args,
|
|
|
|
| 88 |
)
|
| 89 |
|
| 90 |
trainer.train()
|
|
|
|
| 57 |
task_type="CAUSAL_LM",
|
| 58 |
)
|
| 59 |
|
| 60 |
+
# Training config - removed max_seq_length, gradient_checkpointing from SFTConfig
|
| 61 |
training_args = SFTConfig(
|
| 62 |
output_dir="./question-generator-output",
|
| 63 |
num_train_epochs=2,
|
|
|
|
| 73 |
push_to_hub=True,
|
| 74 |
hub_model_id="KevinKeller/cognitive-question-generator-qwen2.5-7b",
|
| 75 |
report_to="none",
|
|
|
|
|
|
|
| 76 |
)
|
| 77 |
|
| 78 |
print("Starting training...")
|
|
|
|
| 83 |
peft_config=peft_config,
|
| 84 |
tokenizer=tokenizer,
|
| 85 |
args=training_args,
|
| 86 |
+
max_seq_length=8192, # Moved here
|
| 87 |
)
|
| 88 |
|
| 89 |
trainer.train()
|