ve-nk-at commited on
Commit
6eab779
·
verified ·
1 Parent(s): f2d9005

Upload folder using huggingface_hub

Browse files
adapter_config.json CHANGED
@@ -30,12 +30,12 @@
30
  "revision": null,
31
  "target_modules": [
32
  "q_proj",
33
- "o_proj",
34
  "up_proj",
35
  "k_proj",
36
- "gate_proj",
37
  "v_proj",
38
- "down_proj"
39
  ],
40
  "target_parameters": null,
41
  "task_type": "SEQ_CLS",
 
30
  "revision": null,
31
  "target_modules": [
32
  "q_proj",
33
+ "gate_proj",
34
  "up_proj",
35
  "k_proj",
36
+ "down_proj",
37
  "v_proj",
38
+ "o_proj"
39
  ],
40
  "target_parameters": null,
41
  "task_type": "SEQ_CLS",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a1c72154a55a5bda67d687cc09755b38f265056bf75adb775d3d2ddfed533e89
3
  size 664593800
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81db12dd49b4fcee7b0227f5f6751db9009649f9c5f037600bba83f05b65099f
3
  size 664593800
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9bc26e086ea3e0e1b228f5203467c3317670ab72c834fe12eccba82ace4d98a6
3
  size 1329396842
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8f596b6105c75cc2ec9f59068bd674b80f36b95967e6aad2bb78f6fe906d794
3
  size 1329396842
trainer_state.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "best_global_step": 80,
3
- "best_metric": 0.6863876863876863,
4
  "best_model_checkpoint": "/content/gemma_lora_imb/checkpoint-80",
5
  "epoch": 1.5686274509803921,
6
  "eval_steps": 20,
@@ -11,66 +11,66 @@
11
  "log_history": [
12
  {
13
  "epoch": 0.39215686274509803,
14
- "grad_norm": 28.07633399963379,
15
  "learning_rate": 8.137254901960784e-06,
16
- "loss": 1.2034,
17
  "step": 20
18
  },
19
  {
20
  "epoch": 0.39215686274509803,
21
- "eval_f1_macro": 0.5441106342393895,
22
- "eval_loss": 0.8645669221878052,
23
- "eval_runtime": 3.5234,
24
- "eval_samples_per_second": 57.616,
25
- "eval_steps_per_second": 1.987,
26
  "step": 20
27
  },
28
  {
29
  "epoch": 0.7843137254901961,
30
- "grad_norm": 33.3344612121582,
31
  "learning_rate": 6.176470588235295e-06,
32
- "loss": 0.7471,
33
  "step": 40
34
  },
35
  {
36
  "epoch": 0.7843137254901961,
37
- "eval_f1_macro": 0.6469418960244648,
38
- "eval_loss": 0.6465137600898743,
39
- "eval_runtime": 3.5045,
40
- "eval_samples_per_second": 57.926,
41
- "eval_steps_per_second": 1.997,
42
  "step": 40
43
  },
44
  {
45
  "epoch": 1.1764705882352942,
46
- "grad_norm": 10.580245018005371,
47
  "learning_rate": 4.215686274509805e-06,
48
- "loss": 0.535,
49
  "step": 60
50
  },
51
  {
52
  "epoch": 1.1764705882352942,
53
- "eval_f1_macro": 0.6720336194020404,
54
- "eval_loss": 0.5972322821617126,
55
- "eval_runtime": 3.5009,
56
- "eval_samples_per_second": 57.986,
57
- "eval_steps_per_second": 2.0,
58
  "step": 60
59
  },
60
  {
61
  "epoch": 1.5686274509803921,
62
- "grad_norm": 7.131758689880371,
63
  "learning_rate": 2.254901960784314e-06,
64
- "loss": 0.3936,
65
  "step": 80
66
  },
67
  {
68
  "epoch": 1.5686274509803921,
69
- "eval_f1_macro": 0.6863876863876863,
70
- "eval_loss": 0.6106746792793274,
71
- "eval_runtime": 3.5083,
72
- "eval_samples_per_second": 57.862,
73
- "eval_steps_per_second": 1.995,
74
  "step": 80
75
  }
76
  ],
 
1
  {
2
  "best_global_step": 80,
3
+ "best_metric": 0.6729883436736228,
4
  "best_model_checkpoint": "/content/gemma_lora_imb/checkpoint-80",
5
  "epoch": 1.5686274509803921,
6
  "eval_steps": 20,
 
11
  "log_history": [
12
  {
13
  "epoch": 0.39215686274509803,
14
+ "grad_norm": 12.482938766479492,
15
  "learning_rate": 8.137254901960784e-06,
16
+ "loss": 1.2092,
17
  "step": 20
18
  },
19
  {
20
  "epoch": 0.39215686274509803,
21
+ "eval_f1_macro": 0.49173730322672554,
22
+ "eval_loss": 0.822236180305481,
23
+ "eval_runtime": 3.4673,
24
+ "eval_samples_per_second": 58.546,
25
+ "eval_steps_per_second": 2.019,
26
  "step": 20
27
  },
28
  {
29
  "epoch": 0.7843137254901961,
30
+ "grad_norm": 34.74234390258789,
31
  "learning_rate": 6.176470588235295e-06,
32
+ "loss": 0.7317,
33
  "step": 40
34
  },
35
  {
36
  "epoch": 0.7843137254901961,
37
+ "eval_f1_macro": 0.6225186035727692,
38
+ "eval_loss": 0.7549954056739807,
39
+ "eval_runtime": 3.4442,
40
+ "eval_samples_per_second": 58.939,
41
+ "eval_steps_per_second": 2.032,
42
  "step": 40
43
  },
44
  {
45
  "epoch": 1.1764705882352942,
46
+ "grad_norm": 11.875113487243652,
47
  "learning_rate": 4.215686274509805e-06,
48
+ "loss": 0.5794,
49
  "step": 60
50
  },
51
  {
52
  "epoch": 1.1764705882352942,
53
+ "eval_f1_macro": 0.6456221198156682,
54
+ "eval_loss": 0.6300039887428284,
55
+ "eval_runtime": 3.4467,
56
+ "eval_samples_per_second": 58.897,
57
+ "eval_steps_per_second": 2.031,
58
  "step": 60
59
  },
60
  {
61
  "epoch": 1.5686274509803921,
62
+ "grad_norm": 9.780080795288086,
63
  "learning_rate": 2.254901960784314e-06,
64
+ "loss": 0.4269,
65
  "step": 80
66
  },
67
  {
68
  "epoch": 1.5686274509803921,
69
+ "eval_f1_macro": 0.6729883436736228,
70
+ "eval_loss": 0.6284839510917664,
71
+ "eval_runtime": 3.4549,
72
+ "eval_samples_per_second": 58.757,
73
+ "eval_steps_per_second": 2.026,
74
  "step": 80
75
  }
76
  ],
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6e1f24e23b5a0d792cfe7dc9996cf0cc5c72a0f7bdeda31eae6235cb53be92db
3
  size 5905
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4bb8e2ee3202964b91665e7e4f6b316ca421ef5dd78c1ca86c9b69263842455
3
  size 5905