jbmurel commited on
Commit
bc75444
·
verified ·
1 Parent(s): e870209

Update bert-base card

Browse files
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -6,7 +6,9 @@ pipeline_tag: fill-mask
6
  ---
7
  # Logion base model
8
 
9
- BERT-based model pretrained on largest set of pre-modern Greek to-date (70+ million words). It was introduced in this [paper](https://aclanthology.org/2023.alp-1.20/). This model ignores cases and accents/diacritics.
 
 
10
 
11
  ## How to use
12
 
 
6
  ---
7
  # Logion base model
8
 
9
+ BERT-based model pretrained on largest set of pre-modern Greek to-date. It was introduced in this [paper](https://aclanthology.org/2023.alp-1.20/).
10
+
11
+ The model uses a WordPiece tokenizer (vocab size of 50,000) on a corpus of over 70 million words (over 95 million tokens) of premodern Greek. This model ignores cases and accents/diacritics.
12
 
13
  ## How to use
14