Training in progress, step 120, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 45118424
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:85fd3349fb0b6d9fe33609fd627bbfa198712a033c54895c31d1954fff998d2f
|
| 3 |
size 45118424
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 23159290
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2ef667d6928d98b9252d546a69290d43eb4b3268ece09ea8d4b8985820baaa05
|
| 3 |
size 23159290
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f9fd0216f0d6187ef1dc8511dbb510792521c2906a3f52d90a5f66d2aa05fa7a
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:61c2b4927e3039b26d377375be782c03ce853d193f96b5868ccf559441e84af9
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric": 0.
|
| 3 |
-
"best_model_checkpoint": "miner_id_besimray/checkpoint-
|
| 4 |
-
"epoch": 1.
|
| 5 |
"eval_steps": 20,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -755,6 +755,154 @@
|
|
| 755 |
"eval_samples_per_second": 21.558,
|
| 756 |
"eval_steps_per_second": 2.235,
|
| 757 |
"step": 100
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 758 |
}
|
| 759 |
],
|
| 760 |
"logging_steps": 1,
|
|
@@ -783,7 +931,7 @@
|
|
| 783 |
"attributes": {}
|
| 784 |
}
|
| 785 |
},
|
| 786 |
-
"total_flos": 2.
|
| 787 |
"train_batch_size": 10,
|
| 788 |
"trial_name": null,
|
| 789 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": 0.7635705471038818,
|
| 3 |
+
"best_model_checkpoint": "miner_id_besimray/checkpoint-120",
|
| 4 |
+
"epoch": 1.5434083601286175,
|
| 5 |
"eval_steps": 20,
|
| 6 |
+
"global_step": 120,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 755 |
"eval_samples_per_second": 21.558,
|
| 756 |
"eval_steps_per_second": 2.235,
|
| 757 |
"step": 100
|
| 758 |
+
},
|
| 759 |
+
{
|
| 760 |
+
"epoch": 1.2990353697749195,
|
| 761 |
+
"grad_norm": 0.27163776755332947,
|
| 762 |
+
"learning_rate": 5.4600950026045326e-05,
|
| 763 |
+
"loss": 0.8166,
|
| 764 |
+
"step": 101
|
| 765 |
+
},
|
| 766 |
+
{
|
| 767 |
+
"epoch": 1.3118971061093248,
|
| 768 |
+
"grad_norm": 0.24301646649837494,
|
| 769 |
+
"learning_rate": 5.261313375270014e-05,
|
| 770 |
+
"loss": 0.7311,
|
| 771 |
+
"step": 102
|
| 772 |
+
},
|
| 773 |
+
{
|
| 774 |
+
"epoch": 1.32475884244373,
|
| 775 |
+
"grad_norm": 0.2601422369480133,
|
| 776 |
+
"learning_rate": 5.0649178193565314e-05,
|
| 777 |
+
"loss": 0.8192,
|
| 778 |
+
"step": 103
|
| 779 |
+
},
|
| 780 |
+
{
|
| 781 |
+
"epoch": 1.337620578778135,
|
| 782 |
+
"grad_norm": 0.28303226828575134,
|
| 783 |
+
"learning_rate": 4.87100722594094e-05,
|
| 784 |
+
"loss": 0.7946,
|
| 785 |
+
"step": 104
|
| 786 |
+
},
|
| 787 |
+
{
|
| 788 |
+
"epoch": 1.3504823151125402,
|
| 789 |
+
"grad_norm": 0.25355738401412964,
|
| 790 |
+
"learning_rate": 4.6796792348466356e-05,
|
| 791 |
+
"loss": 0.7246,
|
| 792 |
+
"step": 105
|
| 793 |
+
},
|
| 794 |
+
{
|
| 795 |
+
"epoch": 1.3633440514469453,
|
| 796 |
+
"grad_norm": 0.2478146255016327,
|
| 797 |
+
"learning_rate": 4.491030185478976e-05,
|
| 798 |
+
"loss": 0.8159,
|
| 799 |
+
"step": 106
|
| 800 |
+
},
|
| 801 |
+
{
|
| 802 |
+
"epoch": 1.3762057877813505,
|
| 803 |
+
"grad_norm": 0.22157523036003113,
|
| 804 |
+
"learning_rate": 4.305155068315481e-05,
|
| 805 |
+
"loss": 0.6485,
|
| 806 |
+
"step": 107
|
| 807 |
+
},
|
| 808 |
+
{
|
| 809 |
+
"epoch": 1.3890675241157555,
|
| 810 |
+
"grad_norm": 0.28142282366752625,
|
| 811 |
+
"learning_rate": 4.12214747707527e-05,
|
| 812 |
+
"loss": 0.7363,
|
| 813 |
+
"step": 108
|
| 814 |
+
},
|
| 815 |
+
{
|
| 816 |
+
"epoch": 1.4019292604501608,
|
| 817 |
+
"grad_norm": 0.21578700840473175,
|
| 818 |
+
"learning_rate": 3.942099561591802e-05,
|
| 819 |
+
"loss": 0.6999,
|
| 820 |
+
"step": 109
|
| 821 |
+
},
|
| 822 |
+
{
|
| 823 |
+
"epoch": 1.414790996784566,
|
| 824 |
+
"grad_norm": 0.24640053510665894,
|
| 825 |
+
"learning_rate": 3.7651019814126654e-05,
|
| 826 |
+
"loss": 0.7388,
|
| 827 |
+
"step": 110
|
| 828 |
+
},
|
| 829 |
+
{
|
| 830 |
+
"epoch": 1.427652733118971,
|
| 831 |
+
"grad_norm": 0.21803902089595795,
|
| 832 |
+
"learning_rate": 3.591243860149759e-05,
|
| 833 |
+
"loss": 0.6631,
|
| 834 |
+
"step": 111
|
| 835 |
+
},
|
| 836 |
+
{
|
| 837 |
+
"epoch": 1.4405144694533762,
|
| 838 |
+
"grad_norm": 0.2544996440410614,
|
| 839 |
+
"learning_rate": 3.4206127406028745e-05,
|
| 840 |
+
"loss": 0.6946,
|
| 841 |
+
"step": 112
|
| 842 |
+
},
|
| 843 |
+
{
|
| 844 |
+
"epoch": 1.4533762057877815,
|
| 845 |
+
"grad_norm": 0.25184303522109985,
|
| 846 |
+
"learning_rate": 3.253294540679257e-05,
|
| 847 |
+
"loss": 0.6731,
|
| 848 |
+
"step": 113
|
| 849 |
+
},
|
| 850 |
+
{
|
| 851 |
+
"epoch": 1.4662379421221865,
|
| 852 |
+
"grad_norm": 0.25497835874557495,
|
| 853 |
+
"learning_rate": 3.089373510131354e-05,
|
| 854 |
+
"loss": 0.7296,
|
| 855 |
+
"step": 114
|
| 856 |
+
},
|
| 857 |
+
{
|
| 858 |
+
"epoch": 1.4790996784565915,
|
| 859 |
+
"grad_norm": 0.273090660572052,
|
| 860 |
+
"learning_rate": 2.9289321881345254e-05,
|
| 861 |
+
"loss": 0.7861,
|
| 862 |
+
"step": 115
|
| 863 |
+
},
|
| 864 |
+
{
|
| 865 |
+
"epoch": 1.4919614147909968,
|
| 866 |
+
"grad_norm": 0.278880774974823,
|
| 867 |
+
"learning_rate": 2.7720513617260856e-05,
|
| 868 |
+
"loss": 0.7271,
|
| 869 |
+
"step": 116
|
| 870 |
+
},
|
| 871 |
+
{
|
| 872 |
+
"epoch": 1.504823151125402,
|
| 873 |
+
"grad_norm": 0.2511500120162964,
|
| 874 |
+
"learning_rate": 2.6188100251265945e-05,
|
| 875 |
+
"loss": 0.6752,
|
| 876 |
+
"step": 117
|
| 877 |
+
},
|
| 878 |
+
{
|
| 879 |
+
"epoch": 1.517684887459807,
|
| 880 |
+
"grad_norm": 0.2742422819137573,
|
| 881 |
+
"learning_rate": 2.4692853399638917e-05,
|
| 882 |
+
"loss": 0.7954,
|
| 883 |
+
"step": 118
|
| 884 |
+
},
|
| 885 |
+
{
|
| 886 |
+
"epoch": 1.5305466237942122,
|
| 887 |
+
"grad_norm": 0.274632066488266,
|
| 888 |
+
"learning_rate": 2.323552596419889e-05,
|
| 889 |
+
"loss": 0.8,
|
| 890 |
+
"step": 119
|
| 891 |
+
},
|
| 892 |
+
{
|
| 893 |
+
"epoch": 1.5434083601286175,
|
| 894 |
+
"grad_norm": 0.25790005922317505,
|
| 895 |
+
"learning_rate": 2.181685175319702e-05,
|
| 896 |
+
"loss": 0.8212,
|
| 897 |
+
"step": 120
|
| 898 |
+
},
|
| 899 |
+
{
|
| 900 |
+
"epoch": 1.5434083601286175,
|
| 901 |
+
"eval_loss": 0.7635705471038818,
|
| 902 |
+
"eval_runtime": 5.7735,
|
| 903 |
+
"eval_samples_per_second": 28.406,
|
| 904 |
+
"eval_steps_per_second": 2.945,
|
| 905 |
+
"step": 120
|
| 906 |
}
|
| 907 |
],
|
| 908 |
"logging_steps": 1,
|
|
|
|
| 931 |
"attributes": {}
|
| 932 |
}
|
| 933 |
},
|
| 934 |
+
"total_flos": 2.858372738187264e+16,
|
| 935 |
"train_batch_size": 10,
|
| 936 |
"trial_name": null,
|
| 937 |
"trial_params": null
|