Add link to paper
Browse filesThis PR ensures the model can be viewed at https://huggingface.co/papers/2410.17215.
Feel free to update the other model cards, and add the paper to the collection :)
README.md
CHANGED
@@ -37,4 +37,14 @@ MiniPLM models achieves better performance given the same computation and scales
|
|
37 |
|
38 |
## Citation
|
39 |
|
40 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
37 |
|
38 |
## Citation
|
39 |
|
40 |
+
```bibtex
|
41 |
+
@misc{gu2024miniplmknowledgedistillationpretraining,
|
42 |
+
title={MiniPLM: Knowledge Distillation for Pre-Training Language Models},
|
43 |
+
author={Yuxian Gu and Hao Zhou and Fandong Meng and Jie Zhou and Minlie Huang},
|
44 |
+
year={2024},
|
45 |
+
eprint={2410.17215},
|
46 |
+
archivePrefix={arXiv},
|
47 |
+
primaryClass={cs.CL},
|
48 |
+
url={https://arxiv.org/abs/2410.17215},
|
49 |
+
}
|
50 |
+
```
|