Yuvrajg2107 commited on
Commit
997e5d7
·
verified ·
1 Parent(s): 52d63af

Training in progress, step 3000, checkpoint

Browse files
last-checkpoint/config.json CHANGED
@@ -19,7 +19,7 @@
19
  "num_hidden_layers": 12,
20
  "pad_token_id": 1,
21
  "position_embedding_type": "absolute",
22
- "transformers_version": "4.57.1",
23
  "type_vocab_size": 1,
24
  "use_cache": true,
25
  "vocab_size": 50265
 
19
  "num_hidden_layers": 12,
20
  "pad_token_id": 1,
21
  "position_embedding_type": "absolute",
22
+ "transformers_version": "4.57.3",
23
  "type_vocab_size": 1,
24
  "use_cache": true,
25
  "vocab_size": 50265
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:79cc233a6be9a69c6d79a84da18b8aa1c085a8ba9ed3fc676b30dd261b8b10c2
3
  size 498612824
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cc173eb2df771065213d42995586028dd871d15a4d0f971a3e924c00d05221b
3
  size 498612824
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2c3f1315d4876eb442bc815dcd6cfd83cf26ce5d2ba594754037d92371255cf1
3
  size 997345931
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14c16fca80bbfab50a9063aa5d2d7ac7fec1181f50a559b2ba121ecfee922439
3
  size 997345931
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8cbb9d20ae0e060525bcac1b70551f336242a3392c3ba0392910b81075e6dc33
3
  size 14645
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abf2b0fc64c4be716a75868bdb69c6ff3620ba2a21085b2c861288da2529d5e7
3
  size 14645
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f4aa03f6e0cd07cf67ce1fbe3101d545f5771ef9148b9debf02b11cf6948da5c
3
  size 1383
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ac1c46a2776d12775d23d0f587efc112188137ce2140da35bc15d301c9f620e
3
  size 1383
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:38a85adeb1e28a9f8d2b9e5eda88bb40a25fff60709bb465ee86ac35b1f82adc
3
  size 1465
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fe4d93dfc6f647f747e92a74d9d5986e59893a476d774e2205030dc67d6ff6d
3
  size 1465
last-checkpoint/trainer_state.json CHANGED
@@ -1,63 +1,86 @@
1
  {
2
  "best_global_step": 1000,
3
- "best_metric": 0.08424179255962372,
4
  "best_model_checkpoint": "./training_output/checkpoint-1000",
5
- "epoch": 0.25,
6
  "eval_steps": 1000,
7
- "global_step": 2000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
- "epoch": 0.0625,
14
- "grad_norm": 33.47822570800781,
15
- "learning_rate": 1.8752500000000004e-05,
16
- "loss": 0.1408,
17
  "step": 500
18
  },
19
  {
20
- "epoch": 0.125,
21
- "grad_norm": 66.1669692993164,
22
- "learning_rate": 1.7502500000000004e-05,
23
- "loss": 0.1244,
24
  "step": 1000
25
  },
26
  {
27
- "epoch": 0.125,
28
- "eval_accuracy": 0.931,
29
- "eval_loss": 0.08424179255962372,
30
- "eval_runtime": 148.5886,
31
- "eval_samples_per_second": 53.84,
32
- "eval_steps_per_second": 3.365,
33
  "step": 1000
34
  },
35
  {
36
- "epoch": 0.1875,
37
- "grad_norm": 40.86865234375,
38
- "learning_rate": 1.62525e-05,
39
- "loss": 0.083,
40
  "step": 1500
41
  },
42
  {
43
- "epoch": 0.25,
44
- "grad_norm": 0.046407029032707214,
45
- "learning_rate": 1.5002500000000002e-05,
46
- "loss": 0.0851,
47
  "step": 2000
48
  },
49
  {
50
- "epoch": 0.25,
51
- "eval_accuracy": 0.7715,
52
- "eval_loss": 0.20412930846214294,
53
- "eval_runtime": 149.0575,
54
- "eval_samples_per_second": 53.671,
55
- "eval_steps_per_second": 3.354,
56
  "step": 2000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
57
  }
58
  ],
59
  "logging_steps": 500,
60
- "max_steps": 8000,
61
  "num_input_tokens_seen": 0,
62
  "num_train_epochs": 1,
63
  "save_steps": 1000,
@@ -73,7 +96,7 @@
73
  "attributes": {}
74
  }
75
  },
76
- "total_flos": 2104888442880000.0,
77
  "train_batch_size": 2,
78
  "trial_name": null,
79
  "trial_params": null
 
1
  {
2
  "best_global_step": 1000,
3
+ "best_metric": 0.10802757740020752,
4
  "best_model_checkpoint": "./training_output/checkpoint-1000",
5
+ "epoch": 0.15,
6
  "eval_steps": 1000,
7
+ "global_step": 3000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
+ "epoch": 0.025,
14
+ "grad_norm": 0.8834348917007446,
15
+ "learning_rate": 1.9501000000000002e-05,
16
+ "loss": 0.1437,
17
  "step": 500
18
  },
19
  {
20
+ "epoch": 0.05,
21
+ "grad_norm": 9.513919830322266,
22
+ "learning_rate": 1.9001e-05,
23
+ "loss": 0.1085,
24
  "step": 1000
25
  },
26
  {
27
+ "epoch": 0.05,
28
+ "eval_accuracy": 0.91835,
29
+ "eval_loss": 0.10802757740020752,
30
+ "eval_runtime": 381.1922,
31
+ "eval_samples_per_second": 52.467,
32
+ "eval_steps_per_second": 3.279,
33
  "step": 1000
34
  },
35
  {
36
+ "epoch": 0.075,
37
+ "grad_norm": 0.013394818641245365,
38
+ "learning_rate": 1.8501e-05,
39
+ "loss": 0.0965,
40
  "step": 1500
41
  },
42
  {
43
+ "epoch": 0.1,
44
+ "grad_norm": 0.017180055379867554,
45
+ "learning_rate": 1.8001000000000003e-05,
46
+ "loss": 0.0716,
47
  "step": 2000
48
  },
49
  {
50
+ "epoch": 0.1,
51
+ "eval_accuracy": 0.8473,
52
+ "eval_loss": 0.25237375497817993,
53
+ "eval_runtime": 381.1616,
54
+ "eval_samples_per_second": 52.471,
55
+ "eval_steps_per_second": 3.279,
56
  "step": 2000
57
+ },
58
+ {
59
+ "epoch": 0.125,
60
+ "grad_norm": 0.017619747668504715,
61
+ "learning_rate": 1.7501e-05,
62
+ "loss": 0.0658,
63
+ "step": 2500
64
+ },
65
+ {
66
+ "epoch": 0.15,
67
+ "grad_norm": 0.527113676071167,
68
+ "learning_rate": 1.7001000000000002e-05,
69
+ "loss": 0.0615,
70
+ "step": 3000
71
+ },
72
+ {
73
+ "epoch": 0.15,
74
+ "eval_accuracy": 0.9299,
75
+ "eval_loss": 0.11818733811378479,
76
+ "eval_runtime": 381.4889,
77
+ "eval_samples_per_second": 52.426,
78
+ "eval_steps_per_second": 3.277,
79
+ "step": 3000
80
  }
81
  ],
82
  "logging_steps": 500,
83
+ "max_steps": 20000,
84
  "num_input_tokens_seen": 0,
85
  "num_train_epochs": 1,
86
  "save_steps": 1000,
 
96
  "attributes": {}
97
  }
98
  },
99
+ "total_flos": 3157332664320000.0,
100
  "train_batch_size": 2,
101
  "trial_name": null,
102
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d1dcf0aa731896a45dc92ce645a0147a079fc2db258f2967bafe27b497eba6b
3
  size 5841
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4672bef08597eaac4166af409003080f023f24990f8f58c185b2b990119dc93e
3
  size 5841