nicholasKluge commited on
Commit
55eaa24
1 Parent(s): c66cfee

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -4
README.md CHANGED
@@ -33,7 +33,7 @@ co2_eq_emissions:
33
  ---
34
  # TeenyTinyLlama-460m-awq
35
 
36
- <img src="./logo.png" alt="A curious llama exploring a mushroom forest." height="200">
37
 
38
  ## Model Summary
39
 
@@ -41,7 +41,7 @@ co2_eq_emissions:
41
 
42
  Large language models (LLMs) have significantly advanced natural language processing, but their progress has yet to be equal across languages. While most LLMs are trained in high-resource languages like English, multilingual models generally underperform monolingual ones. Additionally, aspects of their multilingual foundation sometimes restrict the byproducts they produce, like computational demands and licensing regimes. Hence, we developed the _TeenyTinyLlama_ pair: two compact models for Brazilian Portuguese text generation.
43
 
44
- Read our preprint on [ArXiv](https://arxiv.org/abs/2401.16640).
45
 
46
  ## Details
47
 
@@ -223,7 +223,6 @@ All the shown results are the higher accuracy scores achieved on the respective
223
  ## Cite as 🤗
224
 
225
  ```latex
226
-
227
  @misc{correa24ttllama,
228
  title = {TeenyTinyLlama: open-source tiny language models trained in Brazilian Portuguese},
229
  author = {Corr{\^e}a, Nicholas Kluge and Falk, Sophia and Fatimah, Shiza and Sen, Aniket and De Oliveira, Nythamar},
@@ -231,6 +230,15 @@ All the shown results are the higher accuracy scores achieved on the respective
231
  year={2024}
232
  }
233
 
 
 
 
 
 
 
 
 
 
234
  ```
235
 
236
  ## Funding
@@ -239,4 +247,4 @@ This repository was built as part of the RAIES ([Rede de Inteligência Artificia
239
 
240
  ## License
241
 
242
- TeenyTinyLlama-460m is licensed under the Apache License, Version 2.0. See the [LICENSE](LICENSE) file for more details.
 
33
  ---
34
  # TeenyTinyLlama-460m-awq
35
 
36
+ <img src="../../img/460m-llama.png" alt="A curious llama exploring a mushroom forest." height="200">
37
 
38
  ## Model Summary
39
 
 
41
 
42
  Large language models (LLMs) have significantly advanced natural language processing, but their progress has yet to be equal across languages. While most LLMs are trained in high-resource languages like English, multilingual models generally underperform monolingual ones. Additionally, aspects of their multilingual foundation sometimes restrict the byproducts they produce, like computational demands and licensing regimes. Hence, we developed the _TeenyTinyLlama_ pair: two compact models for Brazilian Portuguese text generation.
43
 
44
+ Read our preprint on [Article](https://www.sciencedirect.com/science/article/pii/S2666827024000343).
45
 
46
  ## Details
47
 
 
223
  ## Cite as 🤗
224
 
225
  ```latex
 
226
  @misc{correa24ttllama,
227
  title = {TeenyTinyLlama: open-source tiny language models trained in Brazilian Portuguese},
228
  author = {Corr{\^e}a, Nicholas Kluge and Falk, Sophia and Fatimah, Shiza and Sen, Aniket and De Oliveira, Nythamar},
 
230
  year={2024}
231
  }
232
 
233
+ @misc{correa24ttllama,
234
+ doi = {10.1016/j.mlwa.2024.100558},
235
+ url = {https://www.sciencedirect.com/science/article/pii/S2666827024000343},
236
+ title = {TeenyTinyLlama: open-source tiny language models trained in Brazilian Portuguese},
237
+ author = {Corr{\^e}a, Nicholas Kluge and Falk, Sophia and Fatimah, Shiza and Sen, Aniket and De Oliveira, Nythamar},
238
+ journal={Machine Learning With Applications},
239
+ publisher = {Springer},
240
+ year={2024}
241
+ }
242
  ```
243
 
244
  ## Funding
 
247
 
248
  ## License
249
 
250
+ TeenyTinyLlama-460m is licensed under the Apache License, Version 2.0. See the [LICENSE](LICENSE) file for more details.