ZYXue commited on
Commit
68f2790
·
verified ·
1 Parent(s): d96c131

Training in progress, step 20

Browse files
README.md CHANGED
@@ -4,8 +4,8 @@ library_name: transformers
4
  model_name: medgemma_lora
5
  tags:
6
  - generated_from_trainer
7
- - trl
8
  - sft
 
9
  licence: license
10
  ---
11
 
 
4
  model_name: medgemma_lora
5
  tags:
6
  - generated_from_trainer
 
7
  - sft
8
+ - trl
9
  licence: license
10
  ---
11
 
adapter_config.json CHANGED
@@ -25,8 +25,8 @@
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
28
- "v_proj",
29
- "q_proj"
30
  ],
31
  "target_parameters": null,
32
  "task_type": "CAUSAL_LM",
 
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
28
+ "q_proj",
29
+ "v_proj"
30
  ],
31
  "target_parameters": null,
32
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e0751ae67e7bdc676d7d24879cb0206affe0e7c4fabce5d3315dbf459e99a055
3
  size 25826856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:043eb1e01faea6da2934972c4f8e8407fbcf31eb2b96afe4cc2a446c7b81fdbf
3
  size 25826856
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b5ba96187501fd47aa8a6fa34ce3ce3b7d1630e332fa48d80537f465630c7f0a
3
- size 6225
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79ea4b5c62dd2fb12515e971775d549f78e488ab8af5dcc925ab3710d08c4764
3
+ size 6289
training_args.json CHANGED
@@ -5,15 +5,15 @@
5
  "lora_target_modules": "q_proj,v_proj",
6
  "tuning_strategy": "lora",
7
  "num_trainable_layers": 5,
8
- "output_dir": "checkpoints/pure/ALL/medgemma_lora",
9
  "num_train_epochs": 3,
10
  "learning_rate": 0.0002,
11
- "per_device_train_batch_size": 1,
12
- "per_device_eval_batch_size": 8,
13
  "gradient_accumulation_steps": 2,
14
  "logging_steps": 10,
15
- "eval_steps": 200,
16
- "save_steps": 2,
17
  "warmup_ratio": 0.03,
18
  "weight_decay": 0.0,
19
  "max_grad_norm": 0.3,
@@ -24,7 +24,7 @@
24
  "gradient_checkpointing": true,
25
  "optim": "adamw_torch_fused",
26
  "ft_type": "SFT",
27
- "data_name": "ALL",
28
  "sft_type": "pure",
29
  "balanced": false
30
  }
 
5
  "lora_target_modules": "q_proj,v_proj",
6
  "tuning_strategy": "lora",
7
  "num_trainable_layers": 5,
8
+ "output_dir": "checkpoints/pure/Chest_xray_pa/medgemma_lora",
9
  "num_train_epochs": 3,
10
  "learning_rate": 0.0002,
11
+ "per_device_train_batch_size": 16,
12
+ "per_device_eval_batch_size": 16,
13
  "gradient_accumulation_steps": 2,
14
  "logging_steps": 10,
15
+ "eval_steps": 20,
16
+ "save_steps": 20,
17
  "warmup_ratio": 0.03,
18
  "weight_decay": 0.0,
19
  "max_grad_norm": 0.3,
 
24
  "gradient_checkpointing": true,
25
  "optim": "adamw_torch_fused",
26
  "ft_type": "SFT",
27
+ "data_name": "Chest_xray_pa",
28
  "sft_type": "pure",
29
  "balanced": false
30
  }