Marijke commited on
Commit
b91375f
·
verified ·
1 Parent(s): 84d1705

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -5
README.md CHANGED
@@ -6,7 +6,7 @@ base_model:
6
 
7
  # Model Card for Model ID
8
 
9
- This model is part of a series of models trained for the ML4AL paper “Gotta catch ‘em all!”: Retrieving people in Ancient Greek texts combining transformer models and domain knowledge written in the context of the KU Leuven ID-N project NIKAW (Networks of Ideas and Knowledge in the Ancient World)
10
 
11
  ## Model Details
12
 
@@ -26,10 +26,26 @@ This model is part of a series of models trained for the ML4AL paper “Gotta ca
26
 
27
  ### Training Data
28
 
29
- **Repository:** [https://github.com/NER-AncientLanguages/NERAncientGreekML4AL] (for data and training scripts)
 
30
 
31
  ### Training Hyperparameters
32
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  ## Evaluation
34
 
35
  This models was evaluation on precision, recall and macro-f1 for its entity classes. See the paper for more information.
@@ -48,6 +64,9 @@ This models was evaluation on precision, recall and macro-f1 for its entity clas
48
 
49
  If you use this work, please cite the following paper:
50
 
 
 
 
51
  ### **BibTeX**
52
  ```bibtex
53
  @inproceedings{Beersmans_Keersmaekers_de Graaf_Van de Cruys_Depauw_Fantoli_2024,
@@ -65,6 +84,3 @@ If you use this work, please cite the following paper:
65
    pages = {152--164}
66
  }
67
 
68
-
69
- **APA:**
70
- Beersmans, M., Keersmaekers, A., de Graaf, E., Van de Cruys, T., Depauw, M., & Fantoli, M. (2024). “Gotta catch `em all!”: Retrieving people in Ancient Greek texts combining transformer models and domain knowledge. In J. Pavlopoulos, T. Sommerschield, Y. Assael, S. Gordin, K. Cho, M. Passarotti, R. Sprugnoli, Y. Liu, B. Li, & A. Anderson (Eds.), Proceedings of the 1st Workshop on Machine Learning for Ancient Languages (ML4AL 2024) (pp. 152–164). Association for Computational Linguistics. https://doi.org/10.18653/v1/2024.ml4al-1.16
 
6
 
7
  # Model Card for Model ID
8
 
9
+ This model is part of a series of models trained for the ML4AL paper “Gotta catch ‘em all!”: Retrieving people in Ancient Greek texts combining transformer models and domain knowledge", written in the context of the KU Leuven ID-N project NIKAW (Networks of Ideas and Knowledge in the Ancient World)
10
 
11
  ## Model Details
12
 
 
26
 
27
  ### Training Data
28
 
29
+ **Repository:** [https://github.com/NER-AncientLanguages/NERAncientGreekML4AL] (for data and training scripts). We thank the following projects for helping to provide the training data:
30
+
31
 
32
  ### Training Hyperparameters
33
 
34
+ We use Weights & Biases for hyperparameter optimization with a random search strategy (10 folds), aiming to maximize the evaluation F1 score (eval_f1).
35
+
36
+ The search space includes:
37
+
38
+ Learning Rate: Sampled uniformly between 1e-6 and 1e-4
39
+ Weight Decay: One of [0.1, 0.01, 0.001]
40
+ Number of Training Epochs: One of [3, 4, 5, 6]
41
+
42
+ For the final training of this model, the hyperparameters were:
43
+ Learning Rate: 9.889410158465026e-05
44
+ Weight Decay: 0.1
45
+ Number of Training Epochs: 5
46
+
47
+
48
+
49
  ## Evaluation
50
 
51
  This models was evaluation on precision, recall and macro-f1 for its entity classes. See the paper for more information.
 
64
 
65
  If you use this work, please cite the following paper:
66
 
67
+ ### **APA:**
68
+ Beersmans, M., Keersmaekers, A., de Graaf, E., Van de Cruys, T., Depauw, M., & Fantoli, M. (2024). “Gotta catch `em all!”: Retrieving people in Ancient Greek texts combining transformer models and domain knowledge. In J. Pavlopoulos, T. Sommerschield, Y. Assael, S. Gordin, K. Cho, M. Passarotti, R. Sprugnoli, Y. Liu, B. Li, & A. Anderson (Eds.), Proceedings of the 1st Workshop on Machine Learning for Ancient Languages (ML4AL 2024) (pp. 152–164). Association for Computational Linguistics. https://doi.org/10.18653/v1/2024.ml4al-1.16
69
+
70
  ### **BibTeX**
71
  ```bibtex
72
  @inproceedings{Beersmans_Keersmaekers_de Graaf_Van de Cruys_Depauw_Fantoli_2024,
 
84
    pages = {152--164}
85
  }
86