craa commited on
Commit
3b2478c
·
verified ·
1 Parent(s): 16d57a2

Model save

Browse files
Files changed (2) hide show
  1. README.md +94 -94
  2. model.safetensors +1 -1
README.md CHANGED
@@ -16,8 +16,8 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 3.2985
20
- - Accuracy: 0.3949
21
 
22
  ## Model description
23
 
@@ -50,98 +50,98 @@ The following hyperparameters were used during training:
50
 
51
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
52
  |:-------------:|:------:|:-----:|:---------------:|:--------:|
53
- | 5.1035 | 0.1078 | 1000 | 5.0181 | 0.2275 |
54
- | 4.5922 | 0.2156 | 2000 | 4.5143 | 0.2699 |
55
- | 4.311 | 0.3235 | 3000 | 4.2451 | 0.2982 |
56
- | 4.159 | 0.4313 | 4000 | 4.0873 | 0.3130 |
57
- | 4.0472 | 0.5391 | 5000 | 3.9894 | 0.3216 |
58
- | 3.9931 | 0.6469 | 6000 | 3.9174 | 0.3284 |
59
- | 3.9191 | 0.7547 | 7000 | 3.8619 | 0.3334 |
60
- | 3.8542 | 0.8625 | 8000 | 3.8161 | 0.3379 |
61
- | 3.8531 | 0.9704 | 9000 | 3.7784 | 0.3417 |
62
- | 3.7613 | 1.0782 | 10000 | 3.7443 | 0.3452 |
63
- | 3.7685 | 1.1860 | 11000 | 3.7209 | 0.3478 |
64
- | 3.7287 | 1.2938 | 12000 | 3.6943 | 0.3499 |
65
- | 3.7051 | 1.4016 | 13000 | 3.6725 | 0.3525 |
66
- | 3.7032 | 1.5094 | 14000 | 3.6559 | 0.3539 |
67
- | 3.6641 | 1.6173 | 15000 | 3.6342 | 0.3560 |
68
- | 3.6638 | 1.7251 | 16000 | 3.6183 | 0.3577 |
69
- | 3.6368 | 1.8329 | 17000 | 3.6027 | 0.3592 |
70
- | 3.6291 | 1.9407 | 18000 | 3.5892 | 0.3600 |
71
- | 3.5668 | 2.0485 | 19000 | 3.5776 | 0.3624 |
72
- | 3.5744 | 2.1563 | 20000 | 3.5742 | 0.3623 |
73
- | 3.5621 | 2.2642 | 21000 | 3.5594 | 0.3641 |
74
- | 3.5697 | 2.3720 | 22000 | 3.5489 | 0.3652 |
75
- | 3.537 | 2.4798 | 23000 | 3.5380 | 0.3660 |
76
- | 3.5319 | 2.5876 | 24000 | 3.5315 | 0.3672 |
77
- | 3.5265 | 2.6954 | 25000 | 3.5187 | 0.3678 |
78
- | 3.5313 | 2.8032 | 26000 | 3.5109 | 0.3687 |
79
- | 3.5433 | 2.9111 | 27000 | 3.5023 | 0.3697 |
80
- | 3.4434 | 3.0189 | 28000 | 3.4995 | 0.3703 |
81
- | 3.4286 | 3.1267 | 29000 | 3.4940 | 0.3712 |
82
- | 3.4506 | 3.2345 | 30000 | 3.4874 | 0.3719 |
83
- | 3.4563 | 3.3423 | 31000 | 3.4811 | 0.3724 |
84
- | 3.4568 | 3.4501 | 32000 | 3.4738 | 0.3734 |
85
- | 3.4606 | 3.5580 | 33000 | 3.4689 | 0.3738 |
86
- | 3.4583 | 3.6658 | 34000 | 3.4624 | 0.3741 |
87
- | 3.459 | 3.7736 | 35000 | 3.4548 | 0.3754 |
88
- | 3.4485 | 3.8814 | 36000 | 3.4491 | 0.3757 |
89
- | 3.4463 | 3.9892 | 37000 | 3.4432 | 0.3763 |
90
- | 3.3631 | 4.0970 | 38000 | 3.4462 | 0.3767 |
91
- | 3.3635 | 4.2049 | 39000 | 3.4433 | 0.3774 |
92
- | 3.3887 | 4.3127 | 40000 | 3.4382 | 0.3775 |
93
- | 3.392 | 4.4205 | 41000 | 3.4320 | 0.3784 |
94
- | 3.3866 | 4.5283 | 42000 | 3.4265 | 0.3786 |
95
- | 3.3958 | 4.6361 | 43000 | 3.4223 | 0.3792 |
96
- | 3.3875 | 4.7439 | 44000 | 3.4181 | 0.3795 |
97
- | 3.4055 | 4.8518 | 45000 | 3.4115 | 0.3803 |
98
- | 3.3721 | 4.9596 | 46000 | 3.4074 | 0.3807 |
99
- | 3.3139 | 5.0674 | 47000 | 3.4086 | 0.3810 |
100
- | 3.3123 | 5.1752 | 48000 | 3.4110 | 0.3812 |
101
- | 3.3424 | 5.2830 | 49000 | 3.4048 | 0.3816 |
102
- | 3.3402 | 5.3908 | 50000 | 3.4008 | 0.3820 |
103
- | 3.3439 | 5.4987 | 51000 | 3.3965 | 0.3823 |
104
- | 3.3189 | 5.6065 | 52000 | 3.3922 | 0.3827 |
105
- | 3.3341 | 5.7143 | 53000 | 3.3859 | 0.3834 |
106
- | 3.3311 | 5.8221 | 54000 | 3.3810 | 0.3840 |
107
- | 3.3347 | 5.9299 | 55000 | 3.3812 | 0.3841 |
108
- | 3.2357 | 6.0377 | 56000 | 3.3827 | 0.3842 |
109
- | 3.2576 | 6.1456 | 57000 | 3.3831 | 0.3845 |
110
- | 3.2761 | 6.2534 | 58000 | 3.3792 | 0.3849 |
111
- | 3.289 | 6.3612 | 59000 | 3.3749 | 0.3853 |
112
- | 3.2893 | 6.4690 | 60000 | 3.3716 | 0.3856 |
113
- | 3.2819 | 6.5768 | 61000 | 3.3661 | 0.3863 |
114
- | 3.2998 | 6.6846 | 62000 | 3.3610 | 0.3867 |
115
- | 3.2773 | 6.7925 | 63000 | 3.3576 | 0.3869 |
116
- | 3.2882 | 6.9003 | 64000 | 3.3532 | 0.3875 |
117
- | 3.1839 | 7.0081 | 65000 | 3.3544 | 0.3875 |
118
- | 3.2252 | 7.1159 | 66000 | 3.3573 | 0.3878 |
119
- | 3.2308 | 7.2237 | 67000 | 3.3543 | 0.3881 |
120
- | 3.2282 | 7.3315 | 68000 | 3.3525 | 0.3883 |
121
- | 3.2182 | 7.4394 | 69000 | 3.3468 | 0.3886 |
122
- | 3.2238 | 7.5472 | 70000 | 3.3437 | 0.3888 |
123
- | 3.2513 | 7.6550 | 71000 | 3.3401 | 0.3895 |
124
- | 3.2518 | 7.7628 | 72000 | 3.3363 | 0.3900 |
125
- | 3.2309 | 7.8706 | 73000 | 3.3322 | 0.3901 |
126
- | 3.2519 | 7.9784 | 74000 | 3.3286 | 0.3906 |
127
- | 3.1583 | 8.0863 | 75000 | 3.3347 | 0.3907 |
128
- | 3.1571 | 8.1941 | 76000 | 3.3333 | 0.3906 |
129
- | 3.1771 | 8.3019 | 77000 | 3.3296 | 0.3910 |
130
- | 3.1673 | 8.4097 | 78000 | 3.3281 | 0.3912 |
131
- | 3.1806 | 8.5175 | 79000 | 3.3227 | 0.3916 |
132
- | 3.1973 | 8.6253 | 80000 | 3.3202 | 0.3922 |
133
- | 3.1893 | 8.7332 | 81000 | 3.3175 | 0.3925 |
134
- | 3.1843 | 8.8410 | 82000 | 3.3137 | 0.3927 |
135
- | 3.1679 | 8.9488 | 83000 | 3.3105 | 0.3930 |
136
- | 3.1293 | 9.0566 | 84000 | 3.3130 | 0.3930 |
137
- | 3.1214 | 9.1644 | 85000 | 3.3126 | 0.3931 |
138
- | 3.1466 | 9.2722 | 86000 | 3.3102 | 0.3936 |
139
- | 3.1247 | 9.3801 | 87000 | 3.3096 | 0.3936 |
140
- | 3.132 | 9.4879 | 88000 | 3.3066 | 0.3939 |
141
- | 3.1121 | 9.5957 | 89000 | 3.3043 | 0.3943 |
142
- | 3.1384 | 9.7035 | 90000 | 3.3019 | 0.3945 |
143
- | 3.1282 | 9.8113 | 91000 | 3.2998 | 0.3947 |
144
- | 3.1198 | 9.9191 | 92000 | 3.2985 | 0.3949 |
145
 
146
 
147
  ### Framework versions
 
16
 
17
  This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 3.3026
20
+ - Accuracy: 0.3942
21
 
22
  ## Model description
23
 
 
50
 
51
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
52
  |:-------------:|:------:|:-----:|:---------------:|:--------:|
53
+ | 5.1027 | 0.1078 | 1000 | 5.0253 | 0.2270 |
54
+ | 4.5966 | 0.2156 | 2000 | 4.5191 | 0.2694 |
55
+ | 4.3193 | 0.3235 | 3000 | 4.2507 | 0.2971 |
56
+ | 4.1641 | 0.4313 | 4000 | 4.0930 | 0.3124 |
57
+ | 4.0541 | 0.5391 | 5000 | 3.9960 | 0.3209 |
58
+ | 4.0006 | 0.6469 | 6000 | 3.9242 | 0.3277 |
59
+ | 3.9272 | 0.7547 | 7000 | 3.8659 | 0.3329 |
60
+ | 3.8604 | 0.8625 | 8000 | 3.8231 | 0.3370 |
61
+ | 3.8606 | 0.9704 | 9000 | 3.7847 | 0.3405 |
62
+ | 3.7669 | 1.0782 | 10000 | 3.7536 | 0.3440 |
63
+ | 3.7755 | 1.1860 | 11000 | 3.7266 | 0.3468 |
64
+ | 3.7364 | 1.2938 | 12000 | 3.6996 | 0.3489 |
65
+ | 3.7135 | 1.4016 | 13000 | 3.6761 | 0.3513 |
66
+ | 3.7085 | 1.5094 | 14000 | 3.6597 | 0.3533 |
67
+ | 3.6698 | 1.6173 | 15000 | 3.6416 | 0.3552 |
68
+ | 3.6694 | 1.7251 | 16000 | 3.6238 | 0.3571 |
69
+ | 3.6432 | 1.8329 | 17000 | 3.6079 | 0.3585 |
70
+ | 3.6354 | 1.9407 | 18000 | 3.5928 | 0.3595 |
71
+ | 3.5732 | 2.0485 | 19000 | 3.5855 | 0.3613 |
72
+ | 3.5788 | 2.1563 | 20000 | 3.5758 | 0.3621 |
73
+ | 3.5674 | 2.2642 | 21000 | 3.5648 | 0.3634 |
74
+ | 3.5759 | 2.3720 | 22000 | 3.5541 | 0.3642 |
75
+ | 3.5439 | 2.4798 | 23000 | 3.5441 | 0.3654 |
76
+ | 3.5368 | 2.5876 | 24000 | 3.5353 | 0.3664 |
77
+ | 3.5327 | 2.6954 | 25000 | 3.5247 | 0.3671 |
78
+ | 3.5385 | 2.8032 | 26000 | 3.5150 | 0.3682 |
79
+ | 3.5497 | 2.9111 | 27000 | 3.5089 | 0.3690 |
80
+ | 3.4502 | 3.0189 | 28000 | 3.5044 | 0.3698 |
81
+ | 3.4347 | 3.1267 | 29000 | 3.5016 | 0.3704 |
82
+ | 3.4554 | 3.2345 | 30000 | 3.4930 | 0.3712 |
83
+ | 3.4623 | 3.3423 | 31000 | 3.4872 | 0.3717 |
84
+ | 3.4633 | 3.4501 | 32000 | 3.4791 | 0.3728 |
85
+ | 3.4665 | 3.5580 | 33000 | 3.4755 | 0.3732 |
86
+ | 3.4633 | 3.6658 | 34000 | 3.4683 | 0.3734 |
87
+ | 3.4656 | 3.7736 | 35000 | 3.4613 | 0.3744 |
88
+ | 3.4539 | 3.8814 | 36000 | 3.4559 | 0.3748 |
89
+ | 3.4528 | 3.9892 | 37000 | 3.4474 | 0.3758 |
90
+ | 3.3674 | 4.0970 | 38000 | 3.4528 | 0.3760 |
91
+ | 3.369 | 4.2049 | 39000 | 3.4489 | 0.3769 |
92
+ | 3.392 | 4.3127 | 40000 | 3.4412 | 0.3772 |
93
+ | 3.3994 | 4.4205 | 41000 | 3.4375 | 0.3774 |
94
+ | 3.3909 | 4.5283 | 42000 | 3.4333 | 0.3779 |
95
+ | 3.4005 | 4.6361 | 43000 | 3.4265 | 0.3783 |
96
+ | 3.3928 | 4.7439 | 44000 | 3.4222 | 0.3788 |
97
+ | 3.4106 | 4.8518 | 45000 | 3.4176 | 0.3796 |
98
+ | 3.3782 | 4.9596 | 46000 | 3.4130 | 0.3801 |
99
+ | 3.3204 | 5.0674 | 47000 | 3.4155 | 0.3804 |
100
+ | 3.3185 | 5.1752 | 48000 | 3.4171 | 0.3807 |
101
+ | 3.3484 | 5.2830 | 49000 | 3.4122 | 0.3810 |
102
+ | 3.3482 | 5.3908 | 50000 | 3.4060 | 0.3813 |
103
+ | 3.3511 | 5.4987 | 51000 | 3.4023 | 0.3817 |
104
+ | 3.3221 | 5.6065 | 52000 | 3.3966 | 0.3822 |
105
+ | 3.339 | 5.7143 | 53000 | 3.3923 | 0.3826 |
106
+ | 3.3373 | 5.8221 | 54000 | 3.3882 | 0.3830 |
107
+ | 3.3391 | 5.9299 | 55000 | 3.3852 | 0.3835 |
108
+ | 3.2419 | 6.0377 | 56000 | 3.3884 | 0.3835 |
109
+ | 3.2621 | 6.1456 | 57000 | 3.3875 | 0.3836 |
110
+ | 3.2821 | 6.2534 | 58000 | 3.3831 | 0.3844 |
111
+ | 3.2945 | 6.3612 | 59000 | 3.3804 | 0.3844 |
112
+ | 3.2938 | 6.4690 | 60000 | 3.3760 | 0.3850 |
113
+ | 3.2862 | 6.5768 | 61000 | 3.3713 | 0.3856 |
114
+ | 3.3055 | 6.6846 | 62000 | 3.3657 | 0.3858 |
115
+ | 3.2823 | 6.7925 | 63000 | 3.3631 | 0.3859 |
116
+ | 3.2918 | 6.9003 | 64000 | 3.3592 | 0.3866 |
117
+ | 3.187 | 7.0081 | 65000 | 3.3601 | 0.3866 |
118
+ | 3.2306 | 7.1159 | 66000 | 3.3626 | 0.3869 |
119
+ | 3.2341 | 7.2237 | 67000 | 3.3604 | 0.3871 |
120
+ | 3.2319 | 7.3315 | 68000 | 3.3570 | 0.3874 |
121
+ | 3.224 | 7.4394 | 69000 | 3.3514 | 0.3878 |
122
+ | 3.2291 | 7.5472 | 70000 | 3.3487 | 0.3879 |
123
+ | 3.2551 | 7.6550 | 71000 | 3.3450 | 0.3887 |
124
+ | 3.2557 | 7.7628 | 72000 | 3.3420 | 0.3890 |
125
+ | 3.2361 | 7.8706 | 73000 | 3.3377 | 0.3893 |
126
+ | 3.2561 | 7.9784 | 74000 | 3.3330 | 0.3899 |
127
+ | 3.1611 | 8.0863 | 75000 | 3.3395 | 0.3898 |
128
+ | 3.1624 | 8.1941 | 76000 | 3.3378 | 0.3899 |
129
+ | 3.1808 | 8.3019 | 77000 | 3.3341 | 0.3901 |
130
+ | 3.1735 | 8.4097 | 78000 | 3.3307 | 0.3905 |
131
+ | 3.185 | 8.5175 | 79000 | 3.3277 | 0.3908 |
132
+ | 3.201 | 8.6253 | 80000 | 3.3242 | 0.3914 |
133
+ | 3.1937 | 8.7332 | 81000 | 3.3214 | 0.3917 |
134
+ | 3.1875 | 8.8410 | 82000 | 3.3186 | 0.3917 |
135
+ | 3.171 | 8.9488 | 83000 | 3.3150 | 0.3924 |
136
+ | 3.133 | 9.0566 | 84000 | 3.3170 | 0.3924 |
137
+ | 3.1247 | 9.1644 | 85000 | 3.3170 | 0.3924 |
138
+ | 3.1504 | 9.2722 | 86000 | 3.3144 | 0.3928 |
139
+ | 3.1282 | 9.3801 | 87000 | 3.3134 | 0.3930 |
140
+ | 3.1361 | 9.4879 | 88000 | 3.3104 | 0.3932 |
141
+ | 3.1146 | 9.5957 | 89000 | 3.3085 | 0.3936 |
142
+ | 3.1429 | 9.7035 | 90000 | 3.3053 | 0.3940 |
143
+ | 3.1332 | 9.8113 | 91000 | 3.3040 | 0.3941 |
144
+ | 3.1257 | 9.9191 | 92000 | 3.3026 | 0.3942 |
145
 
146
 
147
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f6932f7cf5ea43aca11e16a7f41b7958769a26b6c0e237c37c4c5ea02bf2426c
3
  size 503128704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd5610b06252657e535b52e575c8cbea5f40f5a70160253cb7f747ef1bdaf1cc
3
  size 503128704