Fill-Mask
Transformers
Safetensors
roberta
OSainz commited on
Commit
ba13b99
·
verified ·
1 Parent(s): 59f5f22

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -3
README.md CHANGED
@@ -7,12 +7,17 @@ license: apache-2.0
7
 
8
  Submitted to LREC 2026
9
 
10
- ## Model Description
 
 
11
 
12
  BERnaT is a family of monolingual Basque encoder-only language models trained to better represent linguistic variation—including standard, dialectal, historical, and informal Basque—rather than focusing solely on standard textual corpora. Models were trained on corpora that combine high-quality standard Basque with varied sources such as social media and historical texts, aiming to enhance robustness and generalization across natural language understanding (NLU) tasks.
13
 
14
- **Model Types**: Encoder-only Transformer models (RoBERTa-style)
15
- **Languages**: Basque (Euskara)
 
 
 
16
 
17
  ## Training Data
18
 
 
7
 
8
  Submitted to LREC 2026
9
 
10
+ ## Model Details
11
+
12
+ ### Model Description
13
 
14
  BERnaT is a family of monolingual Basque encoder-only language models trained to better represent linguistic variation—including standard, dialectal, historical, and informal Basque—rather than focusing solely on standard textual corpora. Models were trained on corpora that combine high-quality standard Basque with varied sources such as social media and historical texts, aiming to enhance robustness and generalization across natural language understanding (NLU) tasks.
15
 
16
+ - **Developed by:** HiTZ Research Center & IXA Research group (University of the Basque Country UPV/EHU)
17
+ - **Funded by:** Ikergaitu and ALIA projects (Basque and Spanish Government)
18
+ - **License:** Apache 2.0
19
+ - **Model Type**: Encoder-only Transformer models (RoBERTa-style)
20
+ - **Languages**: Basque (Euskara)
21
 
22
  ## Training Data
23