Check commited on
Commit
e05ae20
Β·
1 Parent(s): bafdd89

"auto-commit"

Browse files
Files changed (28) hide show
  1. model-bin/finetune/base/{checkpoint-101551 β†’ checkpoint-123203}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-101551 β†’ checkpoint-123203}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-101551 β†’ checkpoint-123203}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-122953 β†’ checkpoint-123203}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-122953 β†’ checkpoint-123203}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-122953 β†’ checkpoint-123203}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-101551 β†’ checkpoint-123203}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-122953 β†’ checkpoint-123203}/trainer_state.json +323 -5
  9. model-bin/finetune/base/{checkpoint-101551 β†’ checkpoint-123203}/training_args.bin +0 -0
  10. model-bin/finetune/base/{checkpoint-122953 β†’ checkpoint-123577}/config.json +0 -0
  11. model-bin/finetune/base/{checkpoint-122953 β†’ checkpoint-123577}/optimizer.pt +1 -1
  12. model-bin/finetune/base/{checkpoint-122953 β†’ checkpoint-123577}/preprocessor_config.json +0 -0
  13. model-bin/finetune/base/{checkpoint-101551 β†’ checkpoint-123577}/pytorch_model.bin +1 -1
  14. model-bin/finetune/base/{checkpoint-101551 β†’ checkpoint-123577}/rng_state.pth +1 -1
  15. model-bin/finetune/base/{checkpoint-101551 β†’ checkpoint-123577}/scaler.pt +1 -1
  16. model-bin/finetune/base/{checkpoint-122953 β†’ checkpoint-123577}/scheduler.pt +1 -1
  17. model-bin/finetune/base/{checkpoint-101551 β†’ checkpoint-123577}/trainer_state.json +0 -0
  18. model-bin/finetune/base/{checkpoint-122953 β†’ checkpoint-123577}/training_args.bin +0 -0
  19. model-bin/finetune/base/log/1630138521.070973/events.out.tfevents.1630138521.86bb0ddabf9b.4092.81 +3 -0
  20. model-bin/finetune/base/log/1630138919.5753512/events.out.tfevents.1630138919.86bb0ddabf9b.4092.83 +3 -0
  21. model-bin/finetune/base/log/1630139317.2985787/events.out.tfevents.1630139317.86bb0ddabf9b.4092.85 +3 -0
  22. model-bin/finetune/base/log/1630139716.4845998/events.out.tfevents.1630139716.86bb0ddabf9b.4092.87 +3 -0
  23. model-bin/finetune/base/log/1630140109.5076451/events.out.tfevents.1630140109.86bb0ddabf9b.4092.89 +3 -0
  24. model-bin/finetune/base/log/events.out.tfevents.1630138521.86bb0ddabf9b.4092.80 +3 -0
  25. model-bin/finetune/base/log/events.out.tfevents.1630138919.86bb0ddabf9b.4092.82 +3 -0
  26. model-bin/finetune/base/log/events.out.tfevents.1630139317.86bb0ddabf9b.4092.84 +3 -0
  27. model-bin/finetune/base/log/events.out.tfevents.1630139716.86bb0ddabf9b.4092.86 +3 -0
  28. model-bin/finetune/base/log/events.out.tfevents.1630140109.86bb0ddabf9b.4092.88 +3 -0
model-bin/finetune/base/{checkpoint-101551 β†’ checkpoint-123203}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-101551 β†’ checkpoint-123203}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:37bebacb4f6508f032983ad02733dbe2de0ac83e9ad9333cacd5b64b35ae9b07
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e6a125c1f22cc48319045fee252e243c04cd00278a7c172ffc782059f67cc26
3
  size 722165393
model-bin/finetune/base/{checkpoint-101551 β†’ checkpoint-123203}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-122953 β†’ checkpoint-123203}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cec3ce895e777b0b559398c7eedd6d3c5a2ee4f564f98d9b959fa449c27bfc00
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4991cf9566d39f391967058c555dd2bd7c744f264326296fe7df7bd860a29de5
3
  size 377909911
model-bin/finetune/base/{checkpoint-122953 β†’ checkpoint-123203}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da69cf5f555ca72838098676ce3c29a68b9bbdaeff35f98b7e6a2d1f21631a42
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32df719a42d7f5b4cb1229df3210e6e2ee9affa77432670bf3159df58a6ba2ee
3
  size 14503
model-bin/finetune/base/{checkpoint-122953 β†’ checkpoint-123203}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca89118762f5fcee3e0fd4bd629fd68bc9636cb4b91f886dcda93a3a4e9aaae4
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6fc1a7a79ad993600c86fc5a033565a49ad8634a8131af2e3278b03b1b2b7fdb
3
  size 559
model-bin/finetune/base/{checkpoint-101551 β†’ checkpoint-123203}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f9aedb3aeabb99bdad5abed5e156624056cde8a5abae2eb1210e926e544f6890
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f010bc16d14c91a80a5757b57e2aa17704d54320c75c008d1d5c00b5edce0e1c
3
  size 623
model-bin/finetune/base/{checkpoint-122953 β†’ checkpoint-123203}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
- "best_metric": 0.1743826049391605,
3
- "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
- "epoch": 982.9960159362549,
5
- "global_step": 122953,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -255798,11 +255798,329 @@
255798
  "eval_steps_per_second": 0.743,
255799
  "eval_wer": 0.17894194687660953,
255800
  "step": 122953
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
255801
  }
255802
  ],
255803
  "max_steps": 625000,
255804
  "num_train_epochs": 5000,
255805
- "total_flos": 3.460155832498785e+20,
255806
  "trial_name": null,
255807
  "trial_params": null
255808
  }
 
1
  {
2
+ "best_metric": 0.1735723097017633,
3
+ "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-123203",
4
+ "epoch": 984.9960159362549,
5
+ "global_step": 123203,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
255798
  "eval_steps_per_second": 0.743,
255799
  "eval_wer": 0.17894194687660953,
255800
  "step": 122953
255801
+ },
255802
+ {
255803
+ "epoch": 983.02,
255804
+ "learning_rate": 8.046378205128206e-06,
255805
+ "loss": 0.3628,
255806
+ "step": 122955
255807
+ },
255808
+ {
255809
+ "epoch": 983.06,
255810
+ "learning_rate": 8.046298076923078e-06,
255811
+ "loss": 0.2413,
255812
+ "step": 122960
255813
+ },
255814
+ {
255815
+ "epoch": 983.1,
255816
+ "learning_rate": 8.046217948717949e-06,
255817
+ "loss": 0.2896,
255818
+ "step": 122965
255819
+ },
255820
+ {
255821
+ "epoch": 983.14,
255822
+ "learning_rate": 8.046137820512822e-06,
255823
+ "loss": 0.3478,
255824
+ "step": 122970
255825
+ },
255826
+ {
255827
+ "epoch": 983.18,
255828
+ "learning_rate": 8.046057692307693e-06,
255829
+ "loss": 0.6208,
255830
+ "step": 122975
255831
+ },
255832
+ {
255833
+ "epoch": 983.22,
255834
+ "learning_rate": 8.045977564102565e-06,
255835
+ "loss": 0.999,
255836
+ "step": 122980
255837
+ },
255838
+ {
255839
+ "epoch": 983.25,
255840
+ "learning_rate": 8.045897435897436e-06,
255841
+ "loss": 0.2883,
255842
+ "step": 122985
255843
+ },
255844
+ {
255845
+ "epoch": 983.29,
255846
+ "learning_rate": 8.045817307692309e-06,
255847
+ "loss": 0.3098,
255848
+ "step": 122990
255849
+ },
255850
+ {
255851
+ "epoch": 983.33,
255852
+ "learning_rate": 8.04573717948718e-06,
255853
+ "loss": 0.3599,
255854
+ "step": 122995
255855
+ },
255856
+ {
255857
+ "epoch": 983.37,
255858
+ "learning_rate": 8.045657051282052e-06,
255859
+ "loss": 0.6877,
255860
+ "step": 123000
255861
+ },
255862
+ {
255863
+ "epoch": 983.41,
255864
+ "learning_rate": 8.045576923076925e-06,
255865
+ "loss": 0.8981,
255866
+ "step": 123005
255867
+ },
255868
+ {
255869
+ "epoch": 983.45,
255870
+ "learning_rate": 8.045496794871796e-06,
255871
+ "loss": 0.2542,
255872
+ "step": 123010
255873
+ },
255874
+ {
255875
+ "epoch": 983.49,
255876
+ "learning_rate": 8.045416666666668e-06,
255877
+ "loss": 0.2911,
255878
+ "step": 123015
255879
+ },
255880
+ {
255881
+ "epoch": 983.53,
255882
+ "learning_rate": 8.045336538461539e-06,
255883
+ "loss": 0.342,
255884
+ "step": 123020
255885
+ },
255886
+ {
255887
+ "epoch": 983.57,
255888
+ "learning_rate": 8.045256410256412e-06,
255889
+ "loss": 0.5957,
255890
+ "step": 123025
255891
+ },
255892
+ {
255893
+ "epoch": 983.61,
255894
+ "learning_rate": 8.045176282051282e-06,
255895
+ "loss": 1.0049,
255896
+ "step": 123030
255897
+ },
255898
+ {
255899
+ "epoch": 983.65,
255900
+ "learning_rate": 8.045096153846155e-06,
255901
+ "loss": 0.316,
255902
+ "step": 123035
255903
+ },
255904
+ {
255905
+ "epoch": 983.69,
255906
+ "learning_rate": 8.045016025641026e-06,
255907
+ "loss": 0.4509,
255908
+ "step": 123040
255909
+ },
255910
+ {
255911
+ "epoch": 983.73,
255912
+ "learning_rate": 8.044935897435897e-06,
255913
+ "loss": 0.3719,
255914
+ "step": 123045
255915
+ },
255916
+ {
255917
+ "epoch": 983.77,
255918
+ "learning_rate": 8.044855769230769e-06,
255919
+ "loss": 0.5582,
255920
+ "step": 123050
255921
+ },
255922
+ {
255923
+ "epoch": 983.81,
255924
+ "learning_rate": 8.044775641025642e-06,
255925
+ "loss": 0.9277,
255926
+ "step": 123055
255927
+ },
255928
+ {
255929
+ "epoch": 983.85,
255930
+ "learning_rate": 8.044695512820513e-06,
255931
+ "loss": 0.2791,
255932
+ "step": 123060
255933
+ },
255934
+ {
255935
+ "epoch": 983.89,
255936
+ "learning_rate": 8.044615384615385e-06,
255937
+ "loss": 0.2759,
255938
+ "step": 123065
255939
+ },
255940
+ {
255941
+ "epoch": 983.93,
255942
+ "learning_rate": 8.044535256410258e-06,
255943
+ "loss": 0.3229,
255944
+ "step": 123070
255945
+ },
255946
+ {
255947
+ "epoch": 983.97,
255948
+ "learning_rate": 8.044455128205129e-06,
255949
+ "loss": 0.6476,
255950
+ "step": 123075
255951
+ },
255952
+ {
255953
+ "epoch": 984.0,
255954
+ "eval_loss": 0.41966208815574646,
255955
+ "eval_runtime": 36.7239,
255956
+ "eval_samples_per_second": 22.901,
255957
+ "eval_steps_per_second": 0.735,
255958
+ "eval_wer": 0.18895051817252956,
255959
+ "step": 123078
255960
+ },
255961
+ {
255962
+ "epoch": 984.02,
255963
+ "learning_rate": 8.044375e-06,
255964
+ "loss": 0.3195,
255965
+ "step": 123080
255966
+ },
255967
+ {
255968
+ "epoch": 984.06,
255969
+ "learning_rate": 8.044294871794872e-06,
255970
+ "loss": 0.2842,
255971
+ "step": 123085
255972
+ },
255973
+ {
255974
+ "epoch": 984.1,
255975
+ "learning_rate": 8.044214743589745e-06,
255976
+ "loss": 0.2885,
255977
+ "step": 123090
255978
+ },
255979
+ {
255980
+ "epoch": 984.14,
255981
+ "learning_rate": 8.044134615384616e-06,
255982
+ "loss": 0.3852,
255983
+ "step": 123095
255984
+ },
255985
+ {
255986
+ "epoch": 984.18,
255987
+ "learning_rate": 8.044054487179488e-06,
255988
+ "loss": 0.6315,
255989
+ "step": 123100
255990
+ },
255991
+ {
255992
+ "epoch": 984.22,
255993
+ "learning_rate": 8.04397435897436e-06,
255994
+ "loss": 1.1457,
255995
+ "step": 123105
255996
+ },
255997
+ {
255998
+ "epoch": 984.25,
255999
+ "learning_rate": 8.043894230769232e-06,
256000
+ "loss": 0.2628,
256001
+ "step": 123110
256002
+ },
256003
+ {
256004
+ "epoch": 984.29,
256005
+ "learning_rate": 8.043814102564103e-06,
256006
+ "loss": 0.2721,
256007
+ "step": 123115
256008
+ },
256009
+ {
256010
+ "epoch": 984.33,
256011
+ "learning_rate": 8.043733974358975e-06,
256012
+ "loss": 0.3565,
256013
+ "step": 123120
256014
+ },
256015
+ {
256016
+ "epoch": 984.37,
256017
+ "learning_rate": 8.043653846153848e-06,
256018
+ "loss": 0.6264,
256019
+ "step": 123125
256020
+ },
256021
+ {
256022
+ "epoch": 984.41,
256023
+ "learning_rate": 8.043573717948719e-06,
256024
+ "loss": 0.8858,
256025
+ "step": 123130
256026
+ },
256027
+ {
256028
+ "epoch": 984.45,
256029
+ "learning_rate": 8.04349358974359e-06,
256030
+ "loss": 0.3829,
256031
+ "step": 123135
256032
+ },
256033
+ {
256034
+ "epoch": 984.49,
256035
+ "learning_rate": 8.043413461538462e-06,
256036
+ "loss": 0.3229,
256037
+ "step": 123140
256038
+ },
256039
+ {
256040
+ "epoch": 984.53,
256041
+ "learning_rate": 8.043333333333335e-06,
256042
+ "loss": 0.325,
256043
+ "step": 123145
256044
+ },
256045
+ {
256046
+ "epoch": 984.57,
256047
+ "learning_rate": 8.043253205128204e-06,
256048
+ "loss": 0.5855,
256049
+ "step": 123150
256050
+ },
256051
+ {
256052
+ "epoch": 984.61,
256053
+ "learning_rate": 8.043173076923078e-06,
256054
+ "loss": 0.9573,
256055
+ "step": 123155
256056
+ },
256057
+ {
256058
+ "epoch": 984.65,
256059
+ "learning_rate": 8.04309294871795e-06,
256060
+ "loss": 0.3209,
256061
+ "step": 123160
256062
+ },
256063
+ {
256064
+ "epoch": 984.69,
256065
+ "learning_rate": 8.04301282051282e-06,
256066
+ "loss": 0.2824,
256067
+ "step": 123165
256068
+ },
256069
+ {
256070
+ "epoch": 984.73,
256071
+ "learning_rate": 8.042932692307693e-06,
256072
+ "loss": 0.3095,
256073
+ "step": 123170
256074
+ },
256075
+ {
256076
+ "epoch": 984.77,
256077
+ "learning_rate": 8.042852564102565e-06,
256078
+ "loss": 0.6674,
256079
+ "step": 123175
256080
+ },
256081
+ {
256082
+ "epoch": 984.81,
256083
+ "learning_rate": 8.042772435897436e-06,
256084
+ "loss": 0.963,
256085
+ "step": 123180
256086
+ },
256087
+ {
256088
+ "epoch": 984.85,
256089
+ "learning_rate": 8.042692307692307e-06,
256090
+ "loss": 0.2698,
256091
+ "step": 123185
256092
+ },
256093
+ {
256094
+ "epoch": 984.89,
256095
+ "learning_rate": 8.04261217948718e-06,
256096
+ "loss": 0.2897,
256097
+ "step": 123190
256098
+ },
256099
+ {
256100
+ "epoch": 984.93,
256101
+ "learning_rate": 8.042532051282052e-06,
256102
+ "loss": 0.3357,
256103
+ "step": 123195
256104
+ },
256105
+ {
256106
+ "epoch": 984.97,
256107
+ "learning_rate": 8.042451923076923e-06,
256108
+ "loss": 0.6557,
256109
+ "step": 123200
256110
+ },
256111
+ {
256112
+ "epoch": 985.0,
256113
+ "eval_loss": 0.4400167167186737,
256114
+ "eval_runtime": 35.9443,
256115
+ "eval_samples_per_second": 23.425,
256116
+ "eval_steps_per_second": 0.751,
256117
+ "eval_wer": 0.1735723097017633,
256118
+ "step": 123203
256119
  }
256120
  ],
256121
  "max_steps": 625000,
256122
  "num_train_epochs": 5000,
256123
+ "total_flos": 3.467197138225218e+20,
256124
  "trial_name": null,
256125
  "trial_params": null
256126
  }
model-bin/finetune/base/{checkpoint-101551 β†’ checkpoint-123203}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/{checkpoint-122953 β†’ checkpoint-123577}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-122953 β†’ checkpoint-123577}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0ec9b4ca716593f4c0045c041f86101934ebc89734754023031ed5465bdabb1c
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6acc0eac329fac04e079ada3a4e4ba51754b947dcadf089d81de73e171291f61
3
  size 722165393
model-bin/finetune/base/{checkpoint-122953 β†’ checkpoint-123577}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-101551 β†’ checkpoint-123577}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:16460f3d58338e554ea6eebad442b763e3217a5b323751c1d347e261ac643823
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf7df59b02cc7ed9e88e2cd669353a23deeef6a74308cc4e7167ad41676ee351
3
  size 377909911
model-bin/finetune/base/{checkpoint-101551 β†’ checkpoint-123577}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:124fa1581c9c18d202b9651757cc21ab9ba1b088ff1768c107bb1b88075b7694
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d886e18d042453eba71f225443945e4c5e10aa1d554d7f2496aa448ac824c95f
3
  size 14503
model-bin/finetune/base/{checkpoint-101551 β†’ checkpoint-123577}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:04906157368cb6c9283ecf41580e18430167db35e6f89a50d853522cc344492d
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:653c9a365fa09a2ba870fc34c869e859ff2c2d0e6cac59c1f2aa8a0082f4764f
3
  size 559
model-bin/finetune/base/{checkpoint-122953 β†’ checkpoint-123577}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3c117c3e6f981903b3cdb3170ac9334c2facfbc6eadbf698128fa29363fffa85
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6782817a4fd5e153056fb3435c568a9bfe0113ce6e0ca3c04dd844ec1e2e1d6
3
  size 623
model-bin/finetune/base/{checkpoint-101551 β†’ checkpoint-123577}/trainer_state.json RENAMED
The diff for this file is too large to render. See raw diff
 
model-bin/finetune/base/{checkpoint-122953 β†’ checkpoint-123577}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630138521.070973/events.out.tfevents.1630138521.86bb0ddabf9b.4092.81 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b20143bc62770c1cd304460209996bc26e487e757161a38bc07e7643f2798052
3
+ size 4194
model-bin/finetune/base/log/1630138919.5753512/events.out.tfevents.1630138919.86bb0ddabf9b.4092.83 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37749ce5828968f101dc51b92f6b7e61fd51095a12eeedbd152d9fb08bbb7126
3
+ size 4194
model-bin/finetune/base/log/1630139317.2985787/events.out.tfevents.1630139317.86bb0ddabf9b.4092.85 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42ab616fadfff5457e703989d11d29520ee5570ac44a56df22beb8ab38c75664
3
+ size 4194
model-bin/finetune/base/log/1630139716.4845998/events.out.tfevents.1630139716.86bb0ddabf9b.4092.87 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aaaea57c346966bd2a66a31e0d7242ae1aed0f92c4e17aba6aec62bb1ea693f4
3
+ size 4194
model-bin/finetune/base/log/1630140109.5076451/events.out.tfevents.1630140109.86bb0ddabf9b.4092.89 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5480c673301a41248d19b42323bc4443178e67311fd59c60aba63f49ffccf89
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630138521.86bb0ddabf9b.4092.80 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:687e38ca19b074de70c600a73e0a3e999c4828f770f634dfec6432b9fb155ef0
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630138919.86bb0ddabf9b.4092.82 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b68ff634efa64c09bf8ef1172adf6b7e6a5fd348dc3e44f39001a45957063c4
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630139317.86bb0ddabf9b.4092.84 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ceeb1d0646ed9a11b74301771ed6d09a775316cf88a66d9b11a9830c8a8dfaa7
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630139716.86bb0ddabf9b.4092.86 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf07ca2b1f725341225c022419f78325e064baa0cc365bf06e5340d930e5f71a
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630140109.86bb0ddabf9b.4092.88 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3563f4f16037c788db6bd61ef98ee366567e6cf143600cbe0e733aa50468ca58
3
+ size 8622