Training in progress, step 18000, checkpoint
Browse files
last-checkpoint/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 517931840
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c92b8ea1ab7aa1c3c704ca60e66275b713cae4225ae135b904f4c11a6b994994
|
| 3 |
size 517931840
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1035661434
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7a3f3c04ed042650af1b9c11df2cc35ac490889b1116ef774fd4222e5f41e410
|
| 3 |
size 1035661434
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e5e2f21ad13dc4eb631067c76b1a1560519d302bc60e4e9cb00bba81ca70a316
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a234e8f8153fe3070553b0b2d9439870baa50cef586f11ec979ecf56399b8c74
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -2,9 +2,9 @@
|
|
| 2 |
"best_global_step": null,
|
| 3 |
"best_metric": null,
|
| 4 |
"best_model_checkpoint": null,
|
| 5 |
-
"epoch": 0.
|
| 6 |
"eval_steps": 500,
|
| 7 |
-
"global_step":
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
@@ -3034,11 +3034,189 @@
|
|
| 3034 |
"eval_steps_per_second": 19.189,
|
| 3035 |
"num_input_tokens_seen": 17825792000,
|
| 3036 |
"step": 17000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3037 |
}
|
| 3038 |
],
|
| 3039 |
"logging_steps": 50,
|
| 3040 |
"max_steps": 200000,
|
| 3041 |
-
"num_input_tokens_seen":
|
| 3042 |
"num_train_epochs": 5,
|
| 3043 |
"save_steps": 1000,
|
| 3044 |
"stateful_callbacks": {
|
|
@@ -3053,7 +3231,7 @@
|
|
| 3053 |
"attributes": {}
|
| 3054 |
}
|
| 3055 |
},
|
| 3056 |
-
"total_flos": 1.
|
| 3057 |
"train_batch_size": 64,
|
| 3058 |
"trial_name": null,
|
| 3059 |
"trial_params": null
|
|
|
|
| 2 |
"best_global_step": null,
|
| 3 |
"best_metric": null,
|
| 4 |
"best_model_checkpoint": null,
|
| 5 |
+
"epoch": 0.3953887783722338,
|
| 6 |
"eval_steps": 500,
|
| 7 |
+
"global_step": 18000,
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
|
|
| 3034 |
"eval_steps_per_second": 19.189,
|
| 3035 |
"num_input_tokens_seen": 17825792000,
|
| 3036 |
"step": 17000
|
| 3037 |
+
},
|
| 3038 |
+
{
|
| 3039 |
+
"epoch": 0.374521037291477,
|
| 3040 |
+
"grad_norm": 0.1361106038093567,
|
| 3041 |
+
"learning_rate": 0.001,
|
| 3042 |
+
"loss": 2.7342,
|
| 3043 |
+
"num_input_tokens_seen": 17878220800,
|
| 3044 |
+
"step": 17050
|
| 3045 |
+
},
|
| 3046 |
+
{
|
| 3047 |
+
"epoch": 0.37561933945362214,
|
| 3048 |
+
"grad_norm": 0.13839572668075562,
|
| 3049 |
+
"learning_rate": 0.001,
|
| 3050 |
+
"loss": 2.7259,
|
| 3051 |
+
"num_input_tokens_seen": 17930649600,
|
| 3052 |
+
"step": 17100
|
| 3053 |
+
},
|
| 3054 |
+
{
|
| 3055 |
+
"epoch": 0.3767176416157672,
|
| 3056 |
+
"grad_norm": 0.13055244088172913,
|
| 3057 |
+
"learning_rate": 0.001,
|
| 3058 |
+
"loss": 2.7306,
|
| 3059 |
+
"num_input_tokens_seen": 17983078400,
|
| 3060 |
+
"step": 17150
|
| 3061 |
+
},
|
| 3062 |
+
{
|
| 3063 |
+
"epoch": 0.37781594377791233,
|
| 3064 |
+
"grad_norm": 0.1444411724805832,
|
| 3065 |
+
"learning_rate": 0.001,
|
| 3066 |
+
"loss": 2.7315,
|
| 3067 |
+
"num_input_tokens_seen": 18035507200,
|
| 3068 |
+
"step": 17200
|
| 3069 |
+
},
|
| 3070 |
+
{
|
| 3071 |
+
"epoch": 0.3789142459400574,
|
| 3072 |
+
"grad_norm": 0.151028573513031,
|
| 3073 |
+
"learning_rate": 0.001,
|
| 3074 |
+
"loss": 2.7211,
|
| 3075 |
+
"num_input_tokens_seen": 18087936000,
|
| 3076 |
+
"step": 17250
|
| 3077 |
+
},
|
| 3078 |
+
{
|
| 3079 |
+
"epoch": 0.38001254810220253,
|
| 3080 |
+
"grad_norm": 0.15638011693954468,
|
| 3081 |
+
"learning_rate": 0.001,
|
| 3082 |
+
"loss": 2.7269,
|
| 3083 |
+
"num_input_tokens_seen": 18140364800,
|
| 3084 |
+
"step": 17300
|
| 3085 |
+
},
|
| 3086 |
+
{
|
| 3087 |
+
"epoch": 0.3811108502643476,
|
| 3088 |
+
"grad_norm": 0.1508658230304718,
|
| 3089 |
+
"learning_rate": 0.001,
|
| 3090 |
+
"loss": 2.7263,
|
| 3091 |
+
"num_input_tokens_seen": 18192793600,
|
| 3092 |
+
"step": 17350
|
| 3093 |
+
},
|
| 3094 |
+
{
|
| 3095 |
+
"epoch": 0.38220915242649267,
|
| 3096 |
+
"grad_norm": 0.13167701661586761,
|
| 3097 |
+
"learning_rate": 0.001,
|
| 3098 |
+
"loss": 2.7296,
|
| 3099 |
+
"num_input_tokens_seen": 18245222400,
|
| 3100 |
+
"step": 17400
|
| 3101 |
+
},
|
| 3102 |
+
{
|
| 3103 |
+
"epoch": 0.3833074545886378,
|
| 3104 |
+
"grad_norm": 0.14609253406524658,
|
| 3105 |
+
"learning_rate": 0.001,
|
| 3106 |
+
"loss": 2.7249,
|
| 3107 |
+
"num_input_tokens_seen": 18297651200,
|
| 3108 |
+
"step": 17450
|
| 3109 |
+
},
|
| 3110 |
+
{
|
| 3111 |
+
"epoch": 0.38440575675078287,
|
| 3112 |
+
"grad_norm": 0.13172782957553864,
|
| 3113 |
+
"learning_rate": 0.001,
|
| 3114 |
+
"loss": 2.7252,
|
| 3115 |
+
"num_input_tokens_seen": 18350080000,
|
| 3116 |
+
"step": 17500
|
| 3117 |
+
},
|
| 3118 |
+
{
|
| 3119 |
+
"epoch": 0.38440575675078287,
|
| 3120 |
+
"eval_loss": 2.630176544189453,
|
| 3121 |
+
"eval_runtime": 66.0667,
|
| 3122 |
+
"eval_samples_per_second": 75.681,
|
| 3123 |
+
"eval_steps_per_second": 18.92,
|
| 3124 |
+
"num_input_tokens_seen": 18350080000,
|
| 3125 |
+
"step": 17500
|
| 3126 |
+
},
|
| 3127 |
+
{
|
| 3128 |
+
"epoch": 0.385504058912928,
|
| 3129 |
+
"grad_norm": 0.149306520819664,
|
| 3130 |
+
"learning_rate": 0.001,
|
| 3131 |
+
"loss": 2.7245,
|
| 3132 |
+
"num_input_tokens_seen": 18402508800,
|
| 3133 |
+
"step": 17550
|
| 3134 |
+
},
|
| 3135 |
+
{
|
| 3136 |
+
"epoch": 0.38660236107507306,
|
| 3137 |
+
"grad_norm": 0.14191772043704987,
|
| 3138 |
+
"learning_rate": 0.001,
|
| 3139 |
+
"loss": 2.7204,
|
| 3140 |
+
"num_input_tokens_seen": 18454937600,
|
| 3141 |
+
"step": 17600
|
| 3142 |
+
},
|
| 3143 |
+
{
|
| 3144 |
+
"epoch": 0.3877006632372182,
|
| 3145 |
+
"grad_norm": 0.13731072843074799,
|
| 3146 |
+
"learning_rate": 0.001,
|
| 3147 |
+
"loss": 2.7243,
|
| 3148 |
+
"num_input_tokens_seen": 18507366400,
|
| 3149 |
+
"step": 17650
|
| 3150 |
+
},
|
| 3151 |
+
{
|
| 3152 |
+
"epoch": 0.38879896539936326,
|
| 3153 |
+
"grad_norm": 0.1466369777917862,
|
| 3154 |
+
"learning_rate": 0.001,
|
| 3155 |
+
"loss": 2.7262,
|
| 3156 |
+
"num_input_tokens_seen": 18559795200,
|
| 3157 |
+
"step": 17700
|
| 3158 |
+
},
|
| 3159 |
+
{
|
| 3160 |
+
"epoch": 0.38989726756150833,
|
| 3161 |
+
"grad_norm": 0.13290658593177795,
|
| 3162 |
+
"learning_rate": 0.001,
|
| 3163 |
+
"loss": 2.7314,
|
| 3164 |
+
"num_input_tokens_seen": 18612224000,
|
| 3165 |
+
"step": 17750
|
| 3166 |
+
},
|
| 3167 |
+
{
|
| 3168 |
+
"epoch": 0.39099556972365346,
|
| 3169 |
+
"grad_norm": 0.13785040378570557,
|
| 3170 |
+
"learning_rate": 0.001,
|
| 3171 |
+
"loss": 2.7252,
|
| 3172 |
+
"num_input_tokens_seen": 18664652800,
|
| 3173 |
+
"step": 17800
|
| 3174 |
+
},
|
| 3175 |
+
{
|
| 3176 |
+
"epoch": 0.39209387188579853,
|
| 3177 |
+
"grad_norm": 0.13384000957012177,
|
| 3178 |
+
"learning_rate": 0.001,
|
| 3179 |
+
"loss": 2.7321,
|
| 3180 |
+
"num_input_tokens_seen": 18717081600,
|
| 3181 |
+
"step": 17850
|
| 3182 |
+
},
|
| 3183 |
+
{
|
| 3184 |
+
"epoch": 0.39319217404794365,
|
| 3185 |
+
"grad_norm": 0.14927875995635986,
|
| 3186 |
+
"learning_rate": 0.001,
|
| 3187 |
+
"loss": 2.7236,
|
| 3188 |
+
"num_input_tokens_seen": 18769510400,
|
| 3189 |
+
"step": 17900
|
| 3190 |
+
},
|
| 3191 |
+
{
|
| 3192 |
+
"epoch": 0.3942904762100887,
|
| 3193 |
+
"grad_norm": 0.13494938611984253,
|
| 3194 |
+
"learning_rate": 0.001,
|
| 3195 |
+
"loss": 2.7234,
|
| 3196 |
+
"num_input_tokens_seen": 18821939200,
|
| 3197 |
+
"step": 17950
|
| 3198 |
+
},
|
| 3199 |
+
{
|
| 3200 |
+
"epoch": 0.3953887783722338,
|
| 3201 |
+
"grad_norm": 0.15054813027381897,
|
| 3202 |
+
"learning_rate": 0.001,
|
| 3203 |
+
"loss": 2.7236,
|
| 3204 |
+
"num_input_tokens_seen": 18874368000,
|
| 3205 |
+
"step": 18000
|
| 3206 |
+
},
|
| 3207 |
+
{
|
| 3208 |
+
"epoch": 0.3953887783722338,
|
| 3209 |
+
"eval_loss": 2.62626051902771,
|
| 3210 |
+
"eval_runtime": 65.3965,
|
| 3211 |
+
"eval_samples_per_second": 76.457,
|
| 3212 |
+
"eval_steps_per_second": 19.114,
|
| 3213 |
+
"num_input_tokens_seen": 18874368000,
|
| 3214 |
+
"step": 18000
|
| 3215 |
}
|
| 3216 |
],
|
| 3217 |
"logging_steps": 50,
|
| 3218 |
"max_steps": 200000,
|
| 3219 |
+
"num_input_tokens_seen": 18874368000,
|
| 3220 |
"num_train_epochs": 5,
|
| 3221 |
"save_steps": 1000,
|
| 3222 |
"stateful_callbacks": {
|
|
|
|
| 3231 |
"attributes": {}
|
| 3232 |
}
|
| 3233 |
},
|
| 3234 |
+
"total_flos": 1.0749090887368704e+19,
|
| 3235 |
"train_batch_size": 64,
|
| 3236 |
"trial_name": null,
|
| 3237 |
"trial_params": null
|