Training in progress, step 20
Browse files- README.md +1 -1
- adapter_config.json +2 -2
- adapter_model.safetensors +1 -1
- training_args.bin +2 -2
- training_args.json +6 -6
README.md
CHANGED
|
@@ -4,8 +4,8 @@ library_name: transformers
|
|
| 4 |
model_name: medgemma_lora
|
| 5 |
tags:
|
| 6 |
- generated_from_trainer
|
| 7 |
-
- trl
|
| 8 |
- sft
|
|
|
|
| 9 |
licence: license
|
| 10 |
---
|
| 11 |
|
|
|
|
| 4 |
model_name: medgemma_lora
|
| 5 |
tags:
|
| 6 |
- generated_from_trainer
|
|
|
|
| 7 |
- sft
|
| 8 |
+
- trl
|
| 9 |
licence: license
|
| 10 |
---
|
| 11 |
|
adapter_config.json
CHANGED
|
@@ -25,8 +25,8 @@
|
|
| 25 |
"rank_pattern": {},
|
| 26 |
"revision": null,
|
| 27 |
"target_modules": [
|
| 28 |
-
"
|
| 29 |
-
"
|
| 30 |
],
|
| 31 |
"target_parameters": null,
|
| 32 |
"task_type": "CAUSAL_LM",
|
|
|
|
| 25 |
"rank_pattern": {},
|
| 26 |
"revision": null,
|
| 27 |
"target_modules": [
|
| 28 |
+
"q_proj",
|
| 29 |
+
"v_proj"
|
| 30 |
],
|
| 31 |
"target_parameters": null,
|
| 32 |
"task_type": "CAUSAL_LM",
|
adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 25826856
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:043eb1e01faea6da2934972c4f8e8407fbcf31eb2b96afe4cc2a446c7b81fdbf
|
| 3 |
size 25826856
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:79ea4b5c62dd2fb12515e971775d549f78e488ab8af5dcc925ab3710d08c4764
|
| 3 |
+
size 6289
|
training_args.json
CHANGED
|
@@ -5,15 +5,15 @@
|
|
| 5 |
"lora_target_modules": "q_proj,v_proj",
|
| 6 |
"tuning_strategy": "lora",
|
| 7 |
"num_trainable_layers": 5,
|
| 8 |
-
"output_dir": "checkpoints/pure/
|
| 9 |
"num_train_epochs": 3,
|
| 10 |
"learning_rate": 0.0002,
|
| 11 |
-
"per_device_train_batch_size":
|
| 12 |
-
"per_device_eval_batch_size":
|
| 13 |
"gradient_accumulation_steps": 2,
|
| 14 |
"logging_steps": 10,
|
| 15 |
-
"eval_steps":
|
| 16 |
-
"save_steps":
|
| 17 |
"warmup_ratio": 0.03,
|
| 18 |
"weight_decay": 0.0,
|
| 19 |
"max_grad_norm": 0.3,
|
|
@@ -24,7 +24,7 @@
|
|
| 24 |
"gradient_checkpointing": true,
|
| 25 |
"optim": "adamw_torch_fused",
|
| 26 |
"ft_type": "SFT",
|
| 27 |
-
"data_name": "
|
| 28 |
"sft_type": "pure",
|
| 29 |
"balanced": false
|
| 30 |
}
|
|
|
|
| 5 |
"lora_target_modules": "q_proj,v_proj",
|
| 6 |
"tuning_strategy": "lora",
|
| 7 |
"num_trainable_layers": 5,
|
| 8 |
+
"output_dir": "checkpoints/pure/Chest_xray_pa/medgemma_lora",
|
| 9 |
"num_train_epochs": 3,
|
| 10 |
"learning_rate": 0.0002,
|
| 11 |
+
"per_device_train_batch_size": 16,
|
| 12 |
+
"per_device_eval_batch_size": 16,
|
| 13 |
"gradient_accumulation_steps": 2,
|
| 14 |
"logging_steps": 10,
|
| 15 |
+
"eval_steps": 20,
|
| 16 |
+
"save_steps": 20,
|
| 17 |
"warmup_ratio": 0.03,
|
| 18 |
"weight_decay": 0.0,
|
| 19 |
"max_grad_norm": 0.3,
|
|
|
|
| 24 |
"gradient_checkpointing": true,
|
| 25 |
"optim": "adamw_torch_fused",
|
| 26 |
"ft_type": "SFT",
|
| 27 |
+
"data_name": "Chest_xray_pa",
|
| 28 |
"sft_type": "pure",
|
| 29 |
"balanced": false
|
| 30 |
}
|