Add link to paper
Browse filesThis PR ensures the model can be viewed at https://huggingface.co/papers/2410.17215.
Feel free to update the other model cards, and add the paper to the collection :)
README.md
CHANGED
|
@@ -37,4 +37,14 @@ MiniPLM models achieves better performance given the same computation and scales
|
|
| 37 |
|
| 38 |
## Citation
|
| 39 |
|
| 40 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 37 |
|
| 38 |
## Citation
|
| 39 |
|
| 40 |
+
```bibtex
|
| 41 |
+
@misc{gu2024miniplmknowledgedistillationpretraining,
|
| 42 |
+
title={MiniPLM: Knowledge Distillation for Pre-Training Language Models},
|
| 43 |
+
author={Yuxian Gu and Hao Zhou and Fandong Meng and Jie Zhou and Minlie Huang},
|
| 44 |
+
year={2024},
|
| 45 |
+
eprint={2410.17215},
|
| 46 |
+
archivePrefix={arXiv},
|
| 47 |
+
primaryClass={cs.CL},
|
| 48 |
+
url={https://arxiv.org/abs/2410.17215},
|
| 49 |
+
}
|
| 50 |
+
```
|