|
|
--- |
|
|
library_name: transformers |
|
|
license: apache-2.0 |
|
|
datasets: |
|
|
- allmalab/DOLLMA |
|
|
- allmalab/aze-books |
|
|
language: |
|
|
- az |
|
|
base_model: |
|
|
- openai-community/gpt2 |
|
|
pipeline_tag: text-generation |
|
|
--- |
|
|
|
|
|
# Azerbaijani GPT2 Model |
|
|
|
|
|
The model is based on the GPT-2 architecture, specifically trained on Azerbaijani text. It serves as one of the first foundational models designed to generate and understand Azerbaijani language content. Built with the autoregressive transformer decoder architecture, the model generates text token by token, predicting the next word based on the input context. |
|
|
|
|
|
- **Developed by :** aLLMA Lab |
|
|
- **Funded by :** PRODATA LLC |
|
|
- **Model type:** Decoder-only foundational LLM |
|
|
- **Language:** Azerbaijani |
|
|
|
|
|
## Uses |
|
|
|
|
|
The model can be directly used for text generation, sentence completion, next token prediction tasks by providing an input prompt. Additionally, it can be fine-tuned on an Azerbaijani instruction dataset to develop an interactive question-answering model. |
|
|
|
|
|
## Training Details |
|
|
|
|
|
```python |
|
|
context_window=1024 |
|
|
stride=512 |
|
|
|
|
|
lr=1e-3 |
|
|
warmup_steps = 10000 |
|
|
weight_decay=0.1, |
|
|
adam_beta1 = 0.9, |
|
|
adam_beta2 = 0.999 |
|
|
batch_size=512 |
|
|
max_steps=178000 |
|
|
``` |