berkamphoon commited on
Commit
a5b047e
·
verified ·
1 Parent(s): 20eba10

Training in progress, epoch 1

Browse files
README.md CHANGED
@@ -4,8 +4,8 @@ library_name: transformers
4
  model_name: medgemma-27b-it-dr7
5
  tags:
6
  - generated_from_trainer
7
- - trl
8
  - sft
 
9
  licence: license
10
  ---
11
 
@@ -27,7 +27,7 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/yoon307-kaist/medgemma-27b-it-dr7-Project/runs/xum7do4o)
31
 
32
 
33
  This model was trained with SFT.
 
4
  model_name: medgemma-27b-it-dr7
5
  tags:
6
  - generated_from_trainer
 
7
  - sft
8
+ - trl
9
  licence: license
10
  ---
11
 
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/yoon307-kaist/medgemma-27b-it-dr7-Project/runs/ysbi2fv0)
31
 
32
 
33
  This model was trained with SFT.
adapter_config.json CHANGED
@@ -28,16 +28,16 @@
28
  "rank_pattern": {},
29
  "revision": null,
30
  "target_modules": [
31
- "o_proj",
32
- "q_proj",
33
  "fc1",
34
  "v_proj",
35
- "out_proj",
36
- "fc2",
37
- "gate_proj",
38
  "down_proj",
39
- "up_proj",
40
- "k_proj"
 
 
 
41
  ],
42
  "task_type": "CAUSAL_LM",
43
  "trainable_token_indices": null,
 
28
  "rank_pattern": {},
29
  "revision": null,
30
  "target_modules": [
31
+ "out_proj",
32
+ "up_proj",
33
  "fc1",
34
  "v_proj",
 
 
 
35
  "down_proj",
36
+ "fc2",
37
+ "o_proj",
38
+ "q_proj",
39
+ "k_proj",
40
+ "gate_proj"
41
  ],
42
  "task_type": "CAUSAL_LM",
43
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3c0250ec29e4f943185fe8811b4cafef65b141075f8b3d1d0f1ff7e9de07fba3
3
  size 6616400680
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8e6259a4f22e06219271b585070fb37ea36fa4894096a0d2235347ea650bb05
3
  size 6616400680
runs/Jul22_21-50-05_meedgxh100a/events.out.tfevents.1753235406.meedgxh100a.3618383.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:deeffefabc11641cb5525834045fda992d597129b2f4c99d661233dc53cd89b5
3
+ size 9917
train_medgemma_ft_copy.py CHANGED
@@ -303,7 +303,7 @@ if __name__ == '__main__':
303
 
304
  training_args = SFTConfig(
305
  output_dir=exp_name,
306
- num_train_epochs= 5, # Number of training epochs
307
  per_device_train_batch_size=2, # Batch size per device during training
308
  per_device_eval_batch_size=4, # Batch size per device during evaluation
309
  gradient_accumulation_steps=8, # Number of steps before performing a backward/update pass
@@ -313,7 +313,7 @@ if __name__ == '__main__':
313
  save_strategy="epoch", # Save checkpoint every epoch
314
  eval_strategy="steps", # Evaluate every `eval_steps`
315
  eval_steps=10000, # Number of steps between evaluations
316
- learning_rate=2e-5, # Learning rate based on QLoRA paper
317
  bf16=True, # Use bfloat16 precision
318
  max_grad_norm=0.3, # Max gradient norm based on QLoRA paper
319
  warmup_ratio=0.03, # Warmup ratio based on QLoRA paper
 
303
 
304
  training_args = SFTConfig(
305
  output_dir=exp_name,
306
+ num_train_epochs= 15, # Number of training epochs
307
  per_device_train_batch_size=2, # Batch size per device during training
308
  per_device_eval_batch_size=4, # Batch size per device during evaluation
309
  gradient_accumulation_steps=8, # Number of steps before performing a backward/update pass
 
313
  save_strategy="epoch", # Save checkpoint every epoch
314
  eval_strategy="steps", # Evaluate every `eval_steps`
315
  eval_steps=10000, # Number of steps between evaluations
316
+ learning_rate=1e-4, # Learning rate based on QLoRA paper
317
  bf16=True, # Use bfloat16 precision
318
  max_grad_norm=0.3, # Max gradient norm based on QLoRA paper
319
  warmup_ratio=0.03, # Warmup ratio based on QLoRA paper
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f169d34baac7f3033f31c79056229210efe2de0c8a6f3d7543ae292a836d1f64
3
  size 5816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d419952f7a391ce7541d8747f554ed112a42a866627d0740f7b0550c01abdd71
3
  size 5816