alinasrullayev commited on
Commit
d223568
·
verified ·
1 Parent(s): e32f2e0

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +47 -25
README.md CHANGED
@@ -1,24 +1,52 @@
1
  ---
2
  library_name: transformers
 
 
3
  license: mit
4
  base_model: deepset/gbert-base
5
- tags:
6
- - generated_from_keras_callback
 
 
 
 
 
7
  model-index:
8
- - name: alinasrullayev/gbert-base-germaner
9
- results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  ---
11
 
12
- <!-- This model card has been generated automatically according to the information Keras had access to. You should
13
- probably proofread and complete it, then remove this comment. -->
14
 
15
- # alinasrullayev/gbert-base-germaner
16
 
17
- This model is a fine-tuned version of [deepset/gbert-base](https://huggingface.co/deepset/gbert-base) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Train Loss: 0.0139
20
- - Validation Loss: 0.0979
21
- - Epoch: 4
 
22
 
23
  ## Model description
24
 
@@ -37,23 +65,17 @@ More information needed
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
- - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'module': 'keras.optimizers.schedules', 'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 6960, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered_name': None}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01}
41
- - training_precision: float32
42
-
43
- ### Training results
44
-
45
- | Train Loss | Validation Loss | Epoch |
46
- |:----------:|:---------------:|:-----:|
47
- | 0.1273 | 0.0780 | 0 |
48
- | 0.0553 | 0.0798 | 1 |
49
- | 0.0327 | 0.0907 | 2 |
50
- | 0.0205 | 0.0914 | 3 |
51
- | 0.0139 | 0.0979 | 4 |
52
-
53
 
54
  ### Framework versions
55
 
56
  - Transformers 4.46.1
57
- - TensorFlow 2.18.0
58
  - Datasets 3.0.2
59
  - Tokenizers 0.20.1
 
1
  ---
2
  library_name: transformers
3
+ language:
4
+ - de
5
  license: mit
6
  base_model: deepset/gbert-base
7
+ datasets:
8
+ - germaner
9
+ metrics:
10
+ - precision
11
+ - recall
12
+ - f1
13
+ - accuracy
14
  model-index:
15
+ - name: gbert-base-germaner
16
+ results:
17
+ - task:
18
+ name: Token Classification
19
+ type: token-classification
20
+ dataset:
21
+ name: germaner
22
+ type: germaner
23
+ args: default
24
+ metrics:
25
+ - name: precision
26
+ type: precision
27
+ value: 0.8494328804686526
28
+ - name: recall
29
+ type: recall
30
+ value: 0.8772042733942592
31
+ - name: f1
32
+ type: f1
33
+ value: 0.863095238095238
34
+ - name: accuracy
35
+ type: accuracy
36
+ value: 0.9774880173169097
37
  ---
38
 
39
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
40
+ should probably proofread and complete it, then remove this comment. -->
41
 
42
+ # gbert-base-germaner
43
 
44
+ This model is a fine-tuned version of [deepset/gbert-base](https://huggingface.co/deepset/gbert-base) on the germaner dataset.
45
  It achieves the following results on the evaluation set:
46
+ - precision: 0.8494
47
+ - recall: 0.8772
48
+ - f1: 0.8631
49
+ - accuracy: 0.9775
50
 
51
  ## Model description
52
 
 
65
  ### Training hyperparameters
66
 
67
  The following hyperparameters were used during training:
68
+ - num_train_epochs: 5
69
+ - train_batch_size: 16
70
+ - eval_batch_size: 32
71
+ - learning_rate: 2e-05
72
+ - weight_decay_rate: 0.01
73
+ - num_warmup_steps: 0
74
+ - fp16: True
 
 
 
 
 
 
75
 
76
  ### Framework versions
77
 
78
  - Transformers 4.46.1
79
+ - Pytorch 2.5.0+cu121
80
  - Datasets 3.0.2
81
  - Tokenizers 0.20.1