| datasets: | |
| - "wikipedia" | |
| tags: | |
| - "arabic" | |
| - "language-model" | |
| # GPT2-Small-Arabic | |
| ## Model description | |
| GPT2 model from Arabic Wikipedia dataset based on gpt2-small (using Fastai2). | |
| ## Intended uses & limitations | |
| #### How to use | |
| An example is provided in this [colab notebook](https://colab.research.google.com/drive/1mRl7c-5v-Klx27EEAEOAbrfkustL4g7a?usp=sharing). | |
| Both text and poetry (fine-tuned model) generation are included. | |
| #### Limitations and bias | |
| GPT2-small-arabic (trained on Arabic Wikipedia) has several limitations in terms of coverage (Arabic Wikipeedia quality, no diacritics) and training performance. | |
| Use as demonstration or proof of concepts but not as production code. | |
| ## Training data | |
| This pretrained model used the Arabic Wikipedia dump (around 900 MB). | |
| ## Training procedure | |
| Training was done using [Fastai2](https://github.com/fastai/fastai2/) library on Kaggle, using free GPU. | |
| ## Eval results | |
| Final perplexity reached was 72.19, loss: 4.28, accuracy: 0.307 | |
| ### BibTeX entry and citation info | |
| ```bibtex | |
| @inproceedings{Abed Khooli, | |
| year={2020} | |
| } | |
| ``` | |