KevinKeller commited on
Commit
116c7be
·
verified ·
1 Parent(s): ddc6839

Upload train_question_generator.py with huggingface_hub

Browse files
Files changed (1) hide show
  1. train_question_generator.py +2 -3
train_question_generator.py CHANGED
@@ -57,7 +57,7 @@ peft_config = LoraConfig(
57
  task_type="CAUSAL_LM",
58
  )
59
 
60
- # Training config - fewer epochs due to larger dataset
61
  training_args = SFTConfig(
62
  output_dir="./question-generator-output",
63
  num_train_epochs=2,
@@ -73,8 +73,6 @@ training_args = SFTConfig(
73
  push_to_hub=True,
74
  hub_model_id="KevinKeller/cognitive-question-generator-qwen2.5-7b",
75
  report_to="none",
76
- max_seq_length=8192,
77
- gradient_checkpointing=True,
78
  )
79
 
80
  print("Starting training...")
@@ -85,6 +83,7 @@ trainer = SFTTrainer(
85
  peft_config=peft_config,
86
  tokenizer=tokenizer,
87
  args=training_args,
 
88
  )
89
 
90
  trainer.train()
 
57
  task_type="CAUSAL_LM",
58
  )
59
 
60
+ # Training config - removed max_seq_length, gradient_checkpointing from SFTConfig
61
  training_args = SFTConfig(
62
  output_dir="./question-generator-output",
63
  num_train_epochs=2,
 
73
  push_to_hub=True,
74
  hub_model_id="KevinKeller/cognitive-question-generator-qwen2.5-7b",
75
  report_to="none",
 
 
76
  )
77
 
78
  print("Starting training...")
 
83
  peft_config=peft_config,
84
  tokenizer=tokenizer,
85
  args=training_args,
86
+ max_seq_length=8192, # Moved here
87
  )
88
 
89
  trainer.train()