File size: 1,244 Bytes
ad20c71
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
---
language: lv
license: apache-2.0
library_name: transformers
pipeline_tag: fill-mask
tags:
  - latvian
  - encoder
  - fill-mask
  - deberta-v3
---

# lv-deberta-base

Latvian DeBERTaV3 encoder model trained with a replaced token detection (RTD) objective, released with the paper **"Pretraining and Benchmarking Modern Encoders for Latvian"**.

For evaluation code and benchmark results, see:
https://github.com/LUMII-AILab/latvian-encoders

## Citation

```bibtex
@inproceedings{znotins-2026-pretraining,
    title = "Pretraining and Benchmarking Modern Encoders for {L}atvian",
    author = "Znotins, Arturs",
    editor = "Hettiarachchi, Hansi  and
      Ranasinghe, Tharindu  and
      Plum, Alistair  and
      Rayson, Paul  and
      Mitkov, Ruslan  and
      Gaber, Mohamed  and
      Premasiri, Damith  and
      Tan, Fiona Anting  and
      Uyangodage, Lasitha",
    booktitle = "Proceedings of the Second Workshop on Language Models for Low-Resource Languages ({L}o{R}es{LM} 2026)",
    month = mar,
    year = "2026",
    address = "Rabat, Morocco",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2026.loreslm-1.40/",
    pages = "461--470",
    ISBN = "979-8-89176-377-7"
}
```