Azrail commited on
Commit
fbd6bd0
·
verified ·
1 Parent(s): d114f88

Training in progress, step 18000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ebeffb2c037b50b65a6c0dee470a06e80bc04cb18fc25d36c6c23ebbfb1bfdb7
3
  size 517931840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c92b8ea1ab7aa1c3c704ca60e66275b713cae4225ae135b904f4c11a6b994994
3
  size 517931840
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:56670a00c0a6655472a5df0bad61f805ef42230ce33d41f768bab9a708635a97
3
  size 1035661434
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a3f3c04ed042650af1b9c11df2cc35ac490889b1116ef774fd4222e5f41e410
3
  size 1035661434
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b383a5a228123d48b81ff62301f8c357c6f3a9cd7484f11e193f37bbe5162530
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5e2f21ad13dc4eb631067c76b1a1560519d302bc60e4e9cb00bba81ca70a316
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a2af187ec456db07cec83217e48a58a7d4609355155eba34a029dc1dd312e2a7
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a234e8f8153fe3070553b0b2d9439870baa50cef586f11ec979ecf56399b8c74
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 0.37342273512933194,
6
  "eval_steps": 500,
7
- "global_step": 17000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -3034,11 +3034,189 @@
3034
  "eval_steps_per_second": 19.189,
3035
  "num_input_tokens_seen": 17825792000,
3036
  "step": 17000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3037
  }
3038
  ],
3039
  "logging_steps": 50,
3040
  "max_steps": 200000,
3041
- "num_input_tokens_seen": 17825792000,
3042
  "num_train_epochs": 5,
3043
  "save_steps": 1000,
3044
  "stateful_callbacks": {
@@ -3053,7 +3231,7 @@
3053
  "attributes": {}
3054
  }
3055
  },
3056
- "total_flos": 1.0151919171403776e+19,
3057
  "train_batch_size": 64,
3058
  "trial_name": null,
3059
  "trial_params": null
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 0.3953887783722338,
6
  "eval_steps": 500,
7
+ "global_step": 18000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
3034
  "eval_steps_per_second": 19.189,
3035
  "num_input_tokens_seen": 17825792000,
3036
  "step": 17000
3037
+ },
3038
+ {
3039
+ "epoch": 0.374521037291477,
3040
+ "grad_norm": 0.1361106038093567,
3041
+ "learning_rate": 0.001,
3042
+ "loss": 2.7342,
3043
+ "num_input_tokens_seen": 17878220800,
3044
+ "step": 17050
3045
+ },
3046
+ {
3047
+ "epoch": 0.37561933945362214,
3048
+ "grad_norm": 0.13839572668075562,
3049
+ "learning_rate": 0.001,
3050
+ "loss": 2.7259,
3051
+ "num_input_tokens_seen": 17930649600,
3052
+ "step": 17100
3053
+ },
3054
+ {
3055
+ "epoch": 0.3767176416157672,
3056
+ "grad_norm": 0.13055244088172913,
3057
+ "learning_rate": 0.001,
3058
+ "loss": 2.7306,
3059
+ "num_input_tokens_seen": 17983078400,
3060
+ "step": 17150
3061
+ },
3062
+ {
3063
+ "epoch": 0.37781594377791233,
3064
+ "grad_norm": 0.1444411724805832,
3065
+ "learning_rate": 0.001,
3066
+ "loss": 2.7315,
3067
+ "num_input_tokens_seen": 18035507200,
3068
+ "step": 17200
3069
+ },
3070
+ {
3071
+ "epoch": 0.3789142459400574,
3072
+ "grad_norm": 0.151028573513031,
3073
+ "learning_rate": 0.001,
3074
+ "loss": 2.7211,
3075
+ "num_input_tokens_seen": 18087936000,
3076
+ "step": 17250
3077
+ },
3078
+ {
3079
+ "epoch": 0.38001254810220253,
3080
+ "grad_norm": 0.15638011693954468,
3081
+ "learning_rate": 0.001,
3082
+ "loss": 2.7269,
3083
+ "num_input_tokens_seen": 18140364800,
3084
+ "step": 17300
3085
+ },
3086
+ {
3087
+ "epoch": 0.3811108502643476,
3088
+ "grad_norm": 0.1508658230304718,
3089
+ "learning_rate": 0.001,
3090
+ "loss": 2.7263,
3091
+ "num_input_tokens_seen": 18192793600,
3092
+ "step": 17350
3093
+ },
3094
+ {
3095
+ "epoch": 0.38220915242649267,
3096
+ "grad_norm": 0.13167701661586761,
3097
+ "learning_rate": 0.001,
3098
+ "loss": 2.7296,
3099
+ "num_input_tokens_seen": 18245222400,
3100
+ "step": 17400
3101
+ },
3102
+ {
3103
+ "epoch": 0.3833074545886378,
3104
+ "grad_norm": 0.14609253406524658,
3105
+ "learning_rate": 0.001,
3106
+ "loss": 2.7249,
3107
+ "num_input_tokens_seen": 18297651200,
3108
+ "step": 17450
3109
+ },
3110
+ {
3111
+ "epoch": 0.38440575675078287,
3112
+ "grad_norm": 0.13172782957553864,
3113
+ "learning_rate": 0.001,
3114
+ "loss": 2.7252,
3115
+ "num_input_tokens_seen": 18350080000,
3116
+ "step": 17500
3117
+ },
3118
+ {
3119
+ "epoch": 0.38440575675078287,
3120
+ "eval_loss": 2.630176544189453,
3121
+ "eval_runtime": 66.0667,
3122
+ "eval_samples_per_second": 75.681,
3123
+ "eval_steps_per_second": 18.92,
3124
+ "num_input_tokens_seen": 18350080000,
3125
+ "step": 17500
3126
+ },
3127
+ {
3128
+ "epoch": 0.385504058912928,
3129
+ "grad_norm": 0.149306520819664,
3130
+ "learning_rate": 0.001,
3131
+ "loss": 2.7245,
3132
+ "num_input_tokens_seen": 18402508800,
3133
+ "step": 17550
3134
+ },
3135
+ {
3136
+ "epoch": 0.38660236107507306,
3137
+ "grad_norm": 0.14191772043704987,
3138
+ "learning_rate": 0.001,
3139
+ "loss": 2.7204,
3140
+ "num_input_tokens_seen": 18454937600,
3141
+ "step": 17600
3142
+ },
3143
+ {
3144
+ "epoch": 0.3877006632372182,
3145
+ "grad_norm": 0.13731072843074799,
3146
+ "learning_rate": 0.001,
3147
+ "loss": 2.7243,
3148
+ "num_input_tokens_seen": 18507366400,
3149
+ "step": 17650
3150
+ },
3151
+ {
3152
+ "epoch": 0.38879896539936326,
3153
+ "grad_norm": 0.1466369777917862,
3154
+ "learning_rate": 0.001,
3155
+ "loss": 2.7262,
3156
+ "num_input_tokens_seen": 18559795200,
3157
+ "step": 17700
3158
+ },
3159
+ {
3160
+ "epoch": 0.38989726756150833,
3161
+ "grad_norm": 0.13290658593177795,
3162
+ "learning_rate": 0.001,
3163
+ "loss": 2.7314,
3164
+ "num_input_tokens_seen": 18612224000,
3165
+ "step": 17750
3166
+ },
3167
+ {
3168
+ "epoch": 0.39099556972365346,
3169
+ "grad_norm": 0.13785040378570557,
3170
+ "learning_rate": 0.001,
3171
+ "loss": 2.7252,
3172
+ "num_input_tokens_seen": 18664652800,
3173
+ "step": 17800
3174
+ },
3175
+ {
3176
+ "epoch": 0.39209387188579853,
3177
+ "grad_norm": 0.13384000957012177,
3178
+ "learning_rate": 0.001,
3179
+ "loss": 2.7321,
3180
+ "num_input_tokens_seen": 18717081600,
3181
+ "step": 17850
3182
+ },
3183
+ {
3184
+ "epoch": 0.39319217404794365,
3185
+ "grad_norm": 0.14927875995635986,
3186
+ "learning_rate": 0.001,
3187
+ "loss": 2.7236,
3188
+ "num_input_tokens_seen": 18769510400,
3189
+ "step": 17900
3190
+ },
3191
+ {
3192
+ "epoch": 0.3942904762100887,
3193
+ "grad_norm": 0.13494938611984253,
3194
+ "learning_rate": 0.001,
3195
+ "loss": 2.7234,
3196
+ "num_input_tokens_seen": 18821939200,
3197
+ "step": 17950
3198
+ },
3199
+ {
3200
+ "epoch": 0.3953887783722338,
3201
+ "grad_norm": 0.15054813027381897,
3202
+ "learning_rate": 0.001,
3203
+ "loss": 2.7236,
3204
+ "num_input_tokens_seen": 18874368000,
3205
+ "step": 18000
3206
+ },
3207
+ {
3208
+ "epoch": 0.3953887783722338,
3209
+ "eval_loss": 2.62626051902771,
3210
+ "eval_runtime": 65.3965,
3211
+ "eval_samples_per_second": 76.457,
3212
+ "eval_steps_per_second": 19.114,
3213
+ "num_input_tokens_seen": 18874368000,
3214
+ "step": 18000
3215
  }
3216
  ],
3217
  "logging_steps": 50,
3218
  "max_steps": 200000,
3219
+ "num_input_tokens_seen": 18874368000,
3220
  "num_train_epochs": 5,
3221
  "save_steps": 1000,
3222
  "stateful_callbacks": {
 
3231
  "attributes": {}
3232
  }
3233
  },
3234
+ "total_flos": 1.0749090887368704e+19,
3235
  "train_batch_size": 64,
3236
  "trial_name": null,
3237
  "trial_params": null