Anwaarma commited on
Commit
3b0dcb6
·
1 Parent(s): 9526e01

End of training

Browse files
README.md CHANGED
@@ -17,9 +17,9 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [Anwaarma/Merged-Server-praj](https://huggingface.co/Anwaarma/Merged-Server-praj) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.7888
21
- - Accuracy: 0.31
22
- - F1: 0.4733
23
 
24
  ## Model description
25
 
@@ -38,7 +38,7 @@ More information needed
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
- - learning_rate: 5e-05
42
  - train_batch_size: 16
43
  - eval_batch_size: 16
44
  - seed: 42
@@ -50,31 +50,65 @@ The following hyperparameters were used during training:
50
 
51
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
52
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
53
- | No log | 0.0 | 50 | 0.5805 | 0.6 | 0.5992 |
54
- | No log | 0.01 | 100 | 0.5687 | 0.65 | 0.6505 |
55
- | No log | 0.01 | 150 | 0.5711 | 0.64 | 0.6404 |
56
- | No log | 0.01 | 200 | 0.5643 | 0.67 | 0.6678 |
57
- | No log | 0.02 | 250 | 0.5757 | 0.64 | 0.6393 |
58
- | No log | 0.02 | 300 | 0.5855 | 0.66 | 0.6517 |
59
- | No log | 0.02 | 350 | 0.5529 | 0.63 | 0.6260 |
60
- | No log | 0.03 | 400 | 0.5646 | 0.62 | 0.6105 |
61
- | No log | 0.03 | 450 | 0.5814 | 0.61 | 0.6087 |
62
- | 0.5986 | 0.03 | 500 | 0.5702 | 0.62 | 0.6129 |
63
- | 0.5986 | 0.04 | 550 | 0.5618 | 0.62 | 0.6205 |
64
- | 0.5986 | 0.04 | 600 | 0.5517 | 0.63 | 0.6306 |
65
- | 0.5986 | 0.04 | 650 | 0.5522 | 0.62 | 0.6192 |
66
- | 0.5986 | 0.05 | 700 | 0.5524 | 0.63 | 0.6288 |
67
- | 0.5986 | 0.05 | 750 | 0.5619 | 0.64 | 0.6404 |
68
- | 0.5986 | 0.05 | 800 | 0.5638 | 0.64 | 0.64 |
69
- | 0.5986 | 0.06 | 850 | 0.5481 | 0.66 | 0.6605 |
70
- | 0.5986 | 0.06 | 900 | 0.5562 | 0.64 | 0.64 |
71
- | 0.5986 | 0.06 | 950 | 0.5502 | 0.64 | 0.6354 |
72
- | 0.5834 | 0.07 | 1000 | 0.5970 | 0.68 | 0.6606 |
73
- | 0.5834 | 0.07 | 1050 | 0.5369 | 0.67 | 0.6702 |
74
- | 0.5834 | 0.07 | 1100 | 0.5970 | 0.6 | 0.5966 |
75
- | 0.5834 | 0.08 | 1150 | 0.5770 | 0.62 | 0.6192 |
76
- | 0.5834 | 0.08 | 1200 | 0.5582 | 0.63 | 0.6306 |
77
- | 0.5834 | 0.09 | 1250 | 0.5505 | 0.66 | 0.66 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
78
 
79
 
80
  ### Framework versions
 
17
 
18
  This model is a fine-tuned version of [Anwaarma/Merged-Server-praj](https://huggingface.co/Anwaarma/Merged-Server-praj) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.5643
21
+ - Accuracy: 0.82
22
+ - F1: 0.9011
23
 
24
  ## Model description
25
 
 
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
+ - learning_rate: 4e-05
42
  - train_batch_size: 16
43
  - eval_batch_size: 16
44
  - seed: 42
 
50
 
51
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
52
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
53
+ | No log | 0.0 | 50 | 0.5790 | 0.6 | 0.5992 |
54
+ | No log | 0.01 | 100 | 0.5691 | 0.65 | 0.6505 |
55
+ | No log | 0.01 | 150 | 0.5678 | 0.65 | 0.6505 |
56
+ | No log | 0.01 | 200 | 0.5621 | 0.68 | 0.6773 |
57
+ | No log | 0.02 | 250 | 0.5666 | 0.63 | 0.6303 |
58
+ | No log | 0.02 | 300 | 0.5721 | 0.65 | 0.6463 |
59
+ | No log | 0.02 | 350 | 0.5533 | 0.63 | 0.6260 |
60
+ | No log | 0.03 | 400 | 0.5614 | 0.62 | 0.6105 |
61
+ | No log | 0.03 | 450 | 0.5756 | 0.62 | 0.6181 |
62
+ | 0.5985 | 0.03 | 500 | 0.5666 | 0.6 | 0.5947 |
63
+ | 0.5985 | 0.04 | 550 | 0.5613 | 0.64 | 0.6406 |
64
+ | 0.5985 | 0.04 | 600 | 0.5541 | 0.63 | 0.6306 |
65
+ | 0.5985 | 0.04 | 650 | 0.5571 | 0.62 | 0.6192 |
66
+ | 0.5985 | 0.05 | 700 | 0.5536 | 0.62 | 0.6192 |
67
+ | 0.5985 | 0.05 | 750 | 0.5614 | 0.63 | 0.6306 |
68
+ | 0.5985 | 0.05 | 800 | 0.5667 | 0.63 | 0.6297 |
69
+ | 0.5985 | 0.06 | 850 | 0.5466 | 0.66 | 0.6600 |
70
+ | 0.5985 | 0.06 | 900 | 0.5532 | 0.66 | 0.6593 |
71
+ | 0.5985 | 0.06 | 950 | 0.5482 | 0.67 | 0.6630 |
72
+ | 0.5855 | 0.07 | 1000 | 0.5837 | 0.63 | 0.6220 |
73
+ | 0.5855 | 0.07 | 1050 | 0.5368 | 0.67 | 0.6705 |
74
+ | 0.5855 | 0.07 | 1100 | 0.5793 | 0.62 | 0.6167 |
75
+ | 0.5855 | 0.08 | 1150 | 0.5694 | 0.63 | 0.6276 |
76
+ | 0.5855 | 0.08 | 1200 | 0.5520 | 0.63 | 0.6306 |
77
+ | 0.5855 | 0.09 | 1250 | 0.5572 | 0.66 | 0.6593 |
78
+ | 0.5855 | 0.09 | 1300 | 0.5706 | 0.62 | 0.6150 |
79
+ | 0.5855 | 0.09 | 1350 | 0.5694 | 0.66 | 0.6593 |
80
+ | 0.5855 | 0.1 | 1400 | 0.5559 | 0.65 | 0.6497 |
81
+ | 0.5855 | 0.1 | 1450 | 0.5515 | 0.67 | 0.6705 |
82
+ | 0.5777 | 0.1 | 1500 | 0.5447 | 0.64 | 0.6393 |
83
+ | 0.5777 | 0.11 | 1550 | 0.5453 | 0.65 | 0.6502 |
84
+ | 0.5777 | 0.11 | 1600 | 0.5575 | 0.64 | 0.6400 |
85
+ | 0.5777 | 0.11 | 1650 | 0.5498 | 0.66 | 0.6584 |
86
+ | 0.5777 | 0.12 | 1700 | 0.5620 | 0.66 | 0.6604 |
87
+ | 0.5777 | 0.12 | 1750 | 0.5734 | 0.67 | 0.6702 |
88
+ | 0.5777 | 0.12 | 1800 | 0.5561 | 0.66 | 0.6593 |
89
+ | 0.5777 | 0.13 | 1850 | 0.5376 | 0.67 | 0.6649 |
90
+ | 0.5777 | 0.13 | 1900 | 0.5652 | 0.65 | 0.6505 |
91
+ | 0.5777 | 0.13 | 1950 | 0.5414 | 0.67 | 0.6689 |
92
+ | 0.575 | 0.14 | 2000 | 0.5340 | 0.67 | 0.6665 |
93
+ | 0.575 | 0.14 | 2050 | 0.5393 | 0.68 | 0.6794 |
94
+ | 0.575 | 0.14 | 2100 | 0.5253 | 0.7 | 0.6994 |
95
+ | 0.575 | 0.15 | 2150 | 0.5334 | 0.69 | 0.6834 |
96
+ | 0.575 | 0.15 | 2200 | 0.5395 | 0.68 | 0.6773 |
97
+ | 0.575 | 0.15 | 2250 | 0.5426 | 0.65 | 0.6446 |
98
+ | 0.575 | 0.16 | 2300 | 0.5523 | 0.64 | 0.6370 |
99
+ | 0.575 | 0.16 | 2350 | 0.5378 | 0.68 | 0.6804 |
100
+ | 0.575 | 0.16 | 2400 | 0.5375 | 0.67 | 0.6649 |
101
+ | 0.575 | 0.17 | 2450 | 0.5378 | 0.68 | 0.6742 |
102
+ | 0.556 | 0.17 | 2500 | 0.5491 | 0.69 | 0.6867 |
103
+ | 0.556 | 0.17 | 2550 | 0.5347 | 0.66 | 0.6517 |
104
+ | 0.556 | 0.18 | 2600 | 0.5325 | 0.69 | 0.6852 |
105
+ | 0.556 | 0.18 | 2650 | 0.5490 | 0.68 | 0.6794 |
106
+ | 0.556 | 0.18 | 2700 | 0.5313 | 0.7 | 0.7005 |
107
+ | 0.556 | 0.19 | 2750 | 0.5451 | 0.65 | 0.6314 |
108
+ | 0.556 | 0.19 | 2800 | 0.5506 | 0.64 | 0.6312 |
109
+ | 0.556 | 0.19 | 2850 | 0.5539 | 0.65 | 0.6497 |
110
+ | 0.556 | 0.2 | 2900 | 0.5601 | 0.66 | 0.6604 |
111
+ | 0.556 | 0.2 | 2950 | 0.5530 | 0.67 | 0.6705 |
112
 
113
 
114
  ### Framework versions
runs/Dec24_16-50-51_86f59ffbd25a/events.out.tfevents.1703436653.86f59ffbd25a.1423.153 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:859e38bc1dc588f55bfff9bbe4945dc6e1f29dad994ed923bc3748a4e815fa00
3
- size 23565
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2fb32cfda3bf47ffd0d37f778c0662e346f0b3a9d6c90c898b0d288df8f6433
3
+ size 27240
runs/Dec24_16-50-51_86f59ffbd25a/events.out.tfevents.1703436653.86f59ffbd25a.1423.154 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8cde4c5956dc55757865475991bfb253163f021bf22f037a68abebbb25b47649
3
- size 23565
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fbbcb486141a5714a88486d1ed7f1acc62d3160edb3dbc54ac13be6ba67855d
3
+ size 27240
runs/Dec24_16-50-51_86f59ffbd25a/events.out.tfevents.1703436720.86f59ffbd25a.1423.155 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49ae1c5f8c87a93000d14a06f42fbd4a816f579656fac433c215306eb0a0e464
3
+ size 826
runs/Dec24_16-50-51_86f59ffbd25a/events.out.tfevents.1703436720.86f59ffbd25a.1423.156 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85d67cb3fb1fd3612c17151a51e7e5e8f6e562fe229077463394c190c1eba8ef
3
+ size 826