File size: 3,362 Bytes
0fe518b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
# Micka-Gen3

**Author**: [Semantika Research](https://semantika.eu)

## Model Description

**Micka Gen3** is a specialized language model based on the [Microsoft RetNet](https://github.com/microsoft/unilm/tree/master/retnet) architecture, fine-tuned for Retrieval-Augmented Generation (RAG) usage in Slovenian Cultural Heritage Domain. 
It leverages an efficient retention mechanism, and should be used as baseline and in combination with the [GAMS](https://huggingface.co/cjvt/GaMS-9B-Instruct) series of models.

A standalone series of models, based on the GaMS model will also be released.

## Training Data

The model was trained from scratch on:
- **GigaFida corpus** (Slovenian)
- **Slovenian Wikipedia**
- **Random subset of 10,000 English Wikipedia articles**

The model underwent **20 epochs** of training on the above datasets.

### Finetuning

The final stage involved finetuning on **10,000 culturally relevant samples** prepared specifically for the **Povejmo Project**, focusing on cultural heritage content.

## Tokenizer

This model uses the following tokenizer:
- **Tokenizer**: [klokedm/micka-32768](https://huggingface.co/klokedm/micka-32768)

The tokenizer shares the same foundational training data, with additional cultural heritage samples included for domain specificity.

## Architecture

The Micka-Gen3 is based on the **Microsoft RetNet** architecture with the following detailed layers:

- **10 decoder layers**, each including:
  - Retention layers (q_proj, k_proj, v_proj, g_proj, out_proj)
  - Feed-forward layers (linear1, linear2)
- Embedding layer (`embedding.weight`)
- Output projection layers (`out.weight`, `out.bias`)

The architecture is optimized for long-context document retrieval and generation tasks in combination with large Generative AI models.

## Usage

Designed specifically for Retrieval-Augmented Generation (RAG), Micka-Gen3 performs well in:
- Generating contextually accurate responses from Cultural Heritage Texts.

## Funding

The development of the Micka Tokenizer was partially funded by the [PoVeJMo project](https://povejmo.si/), which aims to develop large language models for the Slovenian language.
The project PoVeJMo is cofinanced by:
![ARIS](https://www.cjvt.si/povejmo/wp-content/uploads/sites/28/2023/11/ARISLogoSlo_small.jpg)
![NOO](https://www.cjvt.si/povejmo/wp-content/uploads/sites/28/2023/11/NOO_2023_logotip-transparent_povejmo.png)
![NextGenerationEU](https://www.cjvt.si/povejmo/wp-content/uploads/sites/28/2023/11/Financira_Evropska_unija_2023_logotip-transparent_povejmo.png)


## License

This tokenizer is licensed under the [Creative Commons Attribution-ShareAlike 4.0 International License (CC BY-SA 4.0)](https://creativecommons.org/licenses/by-sa/4.0/). This license allows for sharing and adaptation, provided appropriate credit is given and any derivatives are distributed under the same license.

## Citation

Please cite the following if you use **Micka-Gen3**:

```
@misc{micka-gen3,
  author = {Semantika Research},
  title = {Micka-Gen3: A RetNet-based Slovenian Language Model for RAG tasks},
  year = {2024},
  publisher = {Hugging Face},
  url = {https://huggingface.co/klokedm/micka-gen3}
}
```

## Contact
For more information, please contact:
- [Semantika Research](https://semantika.eu)
- [Hugging Face Repository](https://huggingface.co/klokedm/micka-gen3)