nikoprom commited on
Commit
c0f240c
·
verified ·
1 Parent(s): 2611caa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +34 -10
README.md CHANGED
@@ -3,9 +3,9 @@ license: mit
3
  base_model: deepset/gbert-base
4
  tags:
5
  - generated_from_keras_callback
6
- model-index:
7
- - name: journal_identification_german
8
- results: []
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information Keras had access to. You should
@@ -13,9 +13,8 @@ probably proofread and complete it, then remove this comment. -->
13
 
14
  # journal_identification_german
15
 
16
- This model is a fine-tuned version of [deepset/gbert-base](https://huggingface.co/deepset/gbert-base) on an unknown dataset.
17
- It achieves the following results on the evaluation set:
18
-
19
 
20
  ## Model description
21
 
@@ -23,22 +22,47 @@ More information needed
23
 
24
  ## Intended uses & limitations
25
 
26
- More information needed
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
27
 
28
- ## Training and evaluation data
 
 
 
 
29
 
30
  More information needed
31
 
32
  ## Training procedure
33
 
 
 
34
  ### Training hyperparameters
35
 
36
  The following hyperparameters were used during training:
37
- - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'module': 'keras.optimizers.schedules', 'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 4845, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered_name': None}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01}
 
 
38
  - training_precision: float32
39
 
40
  ### Training results
41
 
 
 
42
 
43
 
44
  ### Framework versions
@@ -46,4 +70,4 @@ The following hyperparameters were used during training:
46
  - Transformers 4.32.0
47
  - TensorFlow 2.14.0
48
  - Datasets 2.12.0
49
- - Tokenizers 0.13.3
 
3
  base_model: deepset/gbert-base
4
  tags:
5
  - generated_from_keras_callback
6
+ language:
7
+ - de
8
+ pipeline_tag: token-classification
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information Keras had access to. You should
 
13
 
14
  # journal_identification_german
15
 
16
+ This model is a fine-tuned version of [deepset/gbert-base](https://huggingface.co/deepset/gbert-base) that was trained to identify references to scientific journals in German news coverage.
17
+ It was trained on a dataset of 8082 annotated paragraphs from German print news articles that was created specifically for this task.
 
18
 
19
  ## Model description
20
 
 
22
 
23
  ## Intended uses & limitations
24
 
25
+ ### How to use
26
+
27
+ You can use this model with a Transformers `pipeline` for token classification:
28
+
29
+ ```python
30
+ >>> from transformers import pipeline
31
+ >>> journal_identifier = pipeline('token-classification', model = 'nikoprom/journal_identification_german')
32
+ >>> sentences = ['Die Pflanze sei im Laufe der Zeit unscheinbarer geworden und damit für Menschen schwerer zu finden, berichten die Forscher im Fachmagazin Current Biology.']
33
+ >>> journal_identifier(sentences)
34
+
35
+ [[{'entity': 'J-Start', 'score': np.float32(0.9984914), 'index': 27, 'word': 'Cur', 'start': 138, 'end': 141},
36
+ {'entity': 'J-Start', 'score': np.float32(0.9978611), 'index': 28, 'word': '##rent', 'start': 141, 'end': 145},
37
+ {'entity': 'J-Inner', 'score': np.float32(0.99738055), 'index': 29, 'word': 'Bio', 'start': 146, 'end': 149},
38
+ {'entity': 'J-Inner', 'score': np.float32(0.9970715), 'index': 30, 'word': '##log', 'start': 149, 'end': 152},
39
+ {'entity': 'J-Inner', 'score': np.float32(0.99715745), 'index': 31, 'word': '##y', 'start': 152, 'end': 153}]]
40
+ ```
41
 
42
+ ### Limitations
43
+
44
+
45
+
46
+ ## Training data
47
 
48
  More information needed
49
 
50
  ## Training procedure
51
 
52
+ The model was trained on a single NVIDIA V100 GPU on the [bwUniCluster 2.0](https://wiki.bwhpc.de/e/BwUniCluster2.0) for 15 epochs with a batch size of 16.
53
+
54
  ### Training hyperparameters
55
 
56
  The following hyperparameters were used during training:
57
+ - learning rate: 2e-5
58
+ - weight decay rate: 0.01
59
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
60
  - training_precision: float32
61
 
62
  ### Training results
63
 
64
+ ## Evaluation
65
+
66
 
67
 
68
  ### Framework versions
 
70
  - Transformers 4.32.0
71
  - TensorFlow 2.14.0
72
  - Datasets 2.12.0
73
+ - Tokenizers 0.13.3