Update README.md
Browse files
README.md
CHANGED
|
@@ -7,7 +7,7 @@ language:
|
|
| 7 |
|
| 8 |
# oyo-bert-base
|
| 9 |
|
| 10 |
-
OYO-BERT (or Oyo-
|
| 11 |
|
| 12 |
### Pre-training corpus
|
| 13 |
A mix of WURA, Wikipedia and MT560 Yoruba data
|
|
|
|
| 7 |
|
| 8 |
# oyo-bert-base
|
| 9 |
|
| 10 |
+
OYO-BERT (or Oyo-dialect of Yoruba BERT) was created by pre-training a [BERT model with token dropping](https://aclanthology.org/2022.acl-long.262/) on Yoruba language texts for about 100K steps. It was trained using BERT-base architecture
|
| 11 |
|
| 12 |
### Pre-training corpus
|
| 13 |
A mix of WURA, Wikipedia and MT560 Yoruba data
|