mgh6 commited on
Commit
a131d2f
·
verified ·
1 Parent(s): fb1ea0f

Training in progress, epoch 10, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7b1930aa8d3a968bfd6ecd49bdd6dc92fbbe16fcbf8df5ba323ebb30856c81c8
3
  size 2682482800
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c30855507821eae60b9a8981ec29a51e2006c4710fe70f6eddd878cdf727bc5
3
  size 2682482800
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c86e35f8ee50fb6da6cf3be46516c8727e2ff2afed6dfeb57b6e68f97501943a
3
  size 5365108834
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd9acbde3767dd0193d5328dd03bad7ab2d4f2ea5faf5f9f997bed12d35d2d5f
3
  size 5365108834
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c5b4ef6bfb1501ea1d61f1cecfe9c5f89d66dd2a0b6e81616014abd495ba17d7
3
  size 15006
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e18b7a226bd09a1c0b40f0d1408dd50fa3221531aba57476aa3bd09f3bf91da
3
  size 15006
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d3a2f6b2401b763e8093a671422c0ca2aede809649bace5a55adda737174083e
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14ff11420ab6e05bda6bd86fde16e8ed4af7e505a292540a3a802d06de28458c
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 9.997999636297509,
5
  "eval_steps": 50,
6
- "global_step": 3430,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1027,6 +1027,111 @@
1027
  "eval_samples_per_second": 41.393,
1028
  "eval_steps_per_second": 20.697,
1029
  "step": 3400
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1030
  }
1031
  ],
1032
  "logging_steps": 50,
@@ -1046,7 +1151,7 @@
1046
  "attributes": {}
1047
  }
1048
  },
1049
- "total_flos": 8.934627923491553e+17,
1050
  "train_batch_size": 2,
1051
  "trial_name": null,
1052
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 10.997999636297509,
5
  "eval_steps": 50,
6
+ "global_step": 3773,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1027
  "eval_samples_per_second": 41.393,
1028
  "eval_steps_per_second": 20.697,
1029
  "step": 3400
1030
+ },
1031
+ {
1032
+ "epoch": 10.05819239861793,
1033
+ "grad_norm": 73.24003601074219,
1034
+ "learning_rate": 4.970845481049563e-05,
1035
+ "loss": 0.7873,
1036
+ "step": 3450
1037
+ },
1038
+ {
1039
+ "epoch": 10.05819239861793,
1040
+ "eval_loss": 0.7501546144485474,
1041
+ "eval_runtime": 117.2672,
1042
+ "eval_samples_per_second": 41.154,
1043
+ "eval_steps_per_second": 20.577,
1044
+ "step": 3450
1045
+ },
1046
+ {
1047
+ "epoch": 10.203673395162756,
1048
+ "grad_norm": 64.36954498291016,
1049
+ "learning_rate": 4.89795918367347e-05,
1050
+ "loss": 0.7681,
1051
+ "step": 3500
1052
+ },
1053
+ {
1054
+ "epoch": 10.203673395162756,
1055
+ "eval_loss": 0.7464452385902405,
1056
+ "eval_runtime": 116.6327,
1057
+ "eval_samples_per_second": 41.378,
1058
+ "eval_steps_per_second": 20.689,
1059
+ "step": 3500
1060
+ },
1061
+ {
1062
+ "epoch": 10.349154391707582,
1063
+ "grad_norm": 53.44997024536133,
1064
+ "learning_rate": 4.825072886297377e-05,
1065
+ "loss": 0.7657,
1066
+ "step": 3550
1067
+ },
1068
+ {
1069
+ "epoch": 10.349154391707582,
1070
+ "eval_loss": 0.7468724250793457,
1071
+ "eval_runtime": 116.3252,
1072
+ "eval_samples_per_second": 41.487,
1073
+ "eval_steps_per_second": 20.744,
1074
+ "step": 3550
1075
+ },
1076
+ {
1077
+ "epoch": 10.49463538825241,
1078
+ "grad_norm": 54.961727142333984,
1079
+ "learning_rate": 4.752186588921283e-05,
1080
+ "loss": 0.7746,
1081
+ "step": 3600
1082
+ },
1083
+ {
1084
+ "epoch": 10.49463538825241,
1085
+ "eval_loss": 0.7428314685821533,
1086
+ "eval_runtime": 116.5215,
1087
+ "eval_samples_per_second": 41.417,
1088
+ "eval_steps_per_second": 20.709,
1089
+ "step": 3600
1090
+ },
1091
+ {
1092
+ "epoch": 10.640116384797237,
1093
+ "grad_norm": 72.78599548339844,
1094
+ "learning_rate": 4.6793002915451894e-05,
1095
+ "loss": 0.7666,
1096
+ "step": 3650
1097
+ },
1098
+ {
1099
+ "epoch": 10.640116384797237,
1100
+ "eval_loss": 0.7381331324577332,
1101
+ "eval_runtime": 116.4741,
1102
+ "eval_samples_per_second": 41.434,
1103
+ "eval_steps_per_second": 20.717,
1104
+ "step": 3650
1105
+ },
1106
+ {
1107
+ "epoch": 10.785597381342063,
1108
+ "grad_norm": 57.89088821411133,
1109
+ "learning_rate": 4.6064139941690965e-05,
1110
+ "loss": 0.7685,
1111
+ "step": 3700
1112
+ },
1113
+ {
1114
+ "epoch": 10.785597381342063,
1115
+ "eval_loss": 0.7335684299468994,
1116
+ "eval_runtime": 116.9024,
1117
+ "eval_samples_per_second": 41.282,
1118
+ "eval_steps_per_second": 20.641,
1119
+ "step": 3700
1120
+ },
1121
+ {
1122
+ "epoch": 10.93107837788689,
1123
+ "grad_norm": 63.70279312133789,
1124
+ "learning_rate": 4.533527696793003e-05,
1125
+ "loss": 0.7666,
1126
+ "step": 3750
1127
+ },
1128
+ {
1129
+ "epoch": 10.93107837788689,
1130
+ "eval_loss": 0.7345423102378845,
1131
+ "eval_runtime": 116.4317,
1132
+ "eval_samples_per_second": 41.449,
1133
+ "eval_steps_per_second": 20.725,
1134
+ "step": 3750
1135
  }
1136
  ],
1137
  "logging_steps": 50,
 
1151
  "attributes": {}
1152
  }
1153
  },
1154
+ "total_flos": 9.827787294554194e+17,
1155
  "train_batch_size": 2,
1156
  "trial_name": null,
1157
  "trial_params": null