Add citations
Browse files
README.md
CHANGED
|
@@ -286,3 +286,20 @@ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-le
|
|
| 286 |
|MuSR (0-shot) | 4.79|
|
| 287 |
|MMLU-PRO (5-shot) | 1.53|
|
| 288 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 286 |
|MuSR (0-shot) | 4.79|
|
| 287 |
|MMLU-PRO (5-shot) | 1.53|
|
| 288 |
|
| 289 |
+
## Citation
|
| 290 |
+
If you use MicroLlama in your research or work, please cite the project using the following reference:
|
| 291 |
+
APA:
|
| 292 |
+
```
|
| 293 |
+
Wang, Z. K. (2024). MicroLlama: A 300M-parameter language model trained from scratch. GitHub & Hugging Face. https://github.com/keeeeenw/MicroLlama, https://huggingface.co/keeeeenw/MicroLlama
|
| 294 |
+
```
|
| 295 |
+
BibTeX:
|
| 296 |
+
```
|
| 297 |
+
@misc{wang2024microllama,
|
| 298 |
+
author = {Zixiao Ken Wang},
|
| 299 |
+
title = {MicroLlama: A 300M-parameter language model trained from scratch},
|
| 300 |
+
year = {2024},
|
| 301 |
+
howpublished = {\url{https://github.com/keeeeenw/MicroLlama}, \url{https://huggingface.co/keeeeenw/MicroLlama}},
|
| 302 |
+
note = {GitHub and Hugging Face repositories}
|
| 303 |
+
}
|
| 304 |
+
```
|
| 305 |
+
🙏 Please cite this work if you find it useful.
|