Add link to paper (#1)
Browse files- Add link to paper (e41e6499bd410a0e5a155208c5facfd7d0e2fe00)
Co-authored-by: Niels Rogge <nielsr@users.noreply.huggingface.co>
README.md
CHANGED
|
@@ -37,4 +37,14 @@ MiniPLM models achieves better performance given the same computation and scales
|
|
| 37 |
|
| 38 |
## Citation
|
| 39 |
|
| 40 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 37 |
|
| 38 |
## Citation
|
| 39 |
|
| 40 |
+
```bibtex
|
| 41 |
+
@misc{gu2024miniplmknowledgedistillationpretraining,
|
| 42 |
+
title={MiniPLM: Knowledge Distillation for Pre-Training Language Models},
|
| 43 |
+
author={Yuxian Gu and Hao Zhou and Fandong Meng and Jie Zhou and Minlie Huang},
|
| 44 |
+
year={2024},
|
| 45 |
+
eprint={2410.17215},
|
| 46 |
+
archivePrefix={arXiv},
|
| 47 |
+
primaryClass={cs.CL},
|
| 48 |
+
url={https://arxiv.org/abs/2410.17215},
|
| 49 |
+
}
|
| 50 |
+
```
|