Update README.md
Browse files
README.md
CHANGED
|
@@ -172,3 +172,16 @@ For details, we refer to the paper and to our benchmark [ChocoLlama-Bench](https
|
|
| 172 |
### Compute Infrastructure
|
| 173 |
|
| 174 |
All ChocoLlama models have been trained on the compute cluster provided by the [Flemish Supercomputer Center (VSC)](https://www.vscentrum.be/). We used 8 to 16 NVIDIA A100 GPU's with 80 GB of VRAM.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 172 |
### Compute Infrastructure
|
| 173 |
|
| 174 |
All ChocoLlama models have been trained on the compute cluster provided by the [Flemish Supercomputer Center (VSC)](https://www.vscentrum.be/). We used 8 to 16 NVIDIA A100 GPU's with 80 GB of VRAM.
|
| 175 |
+
|
| 176 |
+
## Citation
|
| 177 |
+
|
| 178 |
+
If you found this useful for your work, kindly cite our paper:
|
| 179 |
+
|
| 180 |
+
```
|
| 181 |
+
@article{meeus2024chocollama,
|
| 182 |
+
title={ChocoLlama: Lessons Learned From Teaching Llamas Dutch},
|
| 183 |
+
author={Meeus, Matthieu and Rath{\'e}, Anthony and Remy, Fran{\c{c}}ois and Delobelle, Pieter and Decorte, Jens-Joris and Demeester, Thomas},
|
| 184 |
+
journal={arXiv preprint arXiv:2412.07633},
|
| 185 |
+
year={2024}
|
| 186 |
+
}
|
| 187 |
+
```
|