selsayed2003 commited on
Commit
d6b022c
·
verified ·
1 Parent(s): 28a8843

Training in progress, step 600, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:726f1d560af359e7b2df6836f9f371eaba4b7263ff655ba94a33e480bee16453
3
  size 37789864
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1be21ed468e01af6983756e26f22015d1f5801cab0f818bf28403454a69a95f4
3
  size 37789864
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:feabb5160bc6a27a5757ab30a314fdec88f354cd29b012eec7fcdf2d95e99b47
3
  size 75739130
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6acd3f058b34556d890cd08c6abd057bc57a8f3fcadadef66a722c8c6f6d532
3
  size 75739130
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:01f2f2f9671a0d8993679216e22d8fd7eacb25068309e3f889f98e617666e158
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0b6f10790bba0f96b0cc2cb9e9d4f04ac87e60f407c5013fc8f3f1738cf5c72
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0b7153c3fc82f3a3ac998a443c5ffefa8487aa2c23d265568bb13e22e83eaa38
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba96516c7a126eb3ad413f66fd7e0a83cd0c8d8402229e83222c4f7cb6f57202
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 400,
3
- "best_metric": 2.6418941020965576,
4
- "best_model_checkpoint": "business_qa_flan_t5_xl_rope_LoRA_model/checkpoint-400",
5
- "epoch": 0.5533938607868569,
6
  "eval_steps": 200,
7
- "global_step": 400,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -80,6 +80,42 @@
80
  "eval_samples_per_second": 2.416,
81
  "eval_steps_per_second": 0.606,
82
  "step": 400
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
83
  }
84
  ],
85
  "logging_steps": 50,
@@ -99,7 +135,7 @@
99
  "attributes": {}
100
  }
101
  },
102
- "total_flos": 5.49203642155008e+16,
103
  "train_batch_size": 1,
104
  "trial_name": null,
105
  "trial_params": null
 
1
  {
2
+ "best_global_step": 600,
3
+ "best_metric": 2.494161605834961,
4
+ "best_model_checkpoint": "business_qa_flan_t5_xl_rope_LoRA_model/checkpoint-600",
5
+ "epoch": 0.8300907911802854,
6
  "eval_steps": 200,
7
+ "global_step": 600,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
80
  "eval_samples_per_second": 2.416,
81
  "eval_steps_per_second": 0.606,
82
  "step": 400
83
+ },
84
+ {
85
+ "epoch": 0.622568093385214,
86
+ "grad_norm": 2.8039212226867676,
87
+ "learning_rate": 4.3875e-05,
88
+ "loss": 2.7732,
89
+ "step": 450
90
+ },
91
+ {
92
+ "epoch": 0.6917423259835711,
93
+ "grad_norm": 2.8921995162963867,
94
+ "learning_rate": 3.7625e-05,
95
+ "loss": 2.7173,
96
+ "step": 500
97
+ },
98
+ {
99
+ "epoch": 0.7609165585819282,
100
+ "grad_norm": 3.0461184978485107,
101
+ "learning_rate": 3.1375e-05,
102
+ "loss": 2.6135,
103
+ "step": 550
104
+ },
105
+ {
106
+ "epoch": 0.8300907911802854,
107
+ "grad_norm": 3.964676856994629,
108
+ "learning_rate": 2.5124999999999997e-05,
109
+ "loss": 2.6357,
110
+ "step": 600
111
+ },
112
+ {
113
+ "epoch": 0.8300907911802854,
114
+ "eval_loss": 2.494161605834961,
115
+ "eval_runtime": 532.8559,
116
+ "eval_samples_per_second": 2.412,
117
+ "eval_steps_per_second": 0.604,
118
+ "step": 600
119
  }
120
  ],
121
  "logging_steps": 50,
 
135
  "attributes": {}
136
  }
137
  },
138
+ "total_flos": 8.23805463232512e+16,
139
  "train_batch_size": 1,
140
  "trial_name": null,
141
  "trial_params": null