nicholasKluge commited on
Commit
0b09544
1 Parent(s): 4cb071d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -15
README.md CHANGED
@@ -144,26 +144,26 @@ trainer.train()
144
 
145
  ## Fine-Tuning Comparisons
146
 
147
- | Models | [Assin2](https://huggingface.co/datasets/assin2)|
148
- |--------------------------------------------------------------------------------------------|-------------------------------------------------|
149
- | [Bert-large-portuguese-cased](https://huggingface.co/neuralmind/bert-base-portuguese-cased)| 88.97 |
150
- | [Bert-base-portuguese-cased](https://huggingface.co/neuralmind/bert-base-portuguese-cased) | 87.45 |
151
- | [Teeny Tiny Llama 460m](https://huggingface.co/nicholasKluge/TeenyTinyLlama-460m) | 86.43 |
152
- | [Gpt2-small-portuguese](https://huggingface.co/pierreguillou/gpt2-small-portuguese) | 86.11 |
153
- | [Teeny Tiny Llama 160m](https://huggingface.co/nicholasKluge/TeenyTinyLlama-160m) | 85.78 |
 
 
 
154
 
155
  ## Cite as 🤗
156
 
157
  ```latex
158
 
159
- @misc{nicholas22llama,
160
- doi = {10.5281/zenodo.6989727},
161
- url = {https://huggingface.co/nicholasKluge/TeenyTinyLlama-160m},
162
- author = {Nicholas Kluge Corrêa},
163
- title = {TeenyTinyLlama},
164
- year = {2023},
165
- publisher = {HuggingFace},
166
- journal = {HuggingFace repository},
167
  }
168
 
169
  ```
 
144
 
145
  ## Fine-Tuning Comparisons
146
 
147
+ To further evaluate the downstream capabilities of our models, we decided to employ a basic fine-tuning procedure for our TTL pair on a subset of tasks from the Poeta benchmark. We apply the same procedure for comparison purposes on both [BERTimbau](https://huggingface.co/neuralmind/bert-base-portuguese-cased) models, given that they are also LLM trained from scratch in Brazilian Portuguese and have a similar size range to our models. We used these comparisons to assess if our pre-training runs produced LLM capable of producing good results ("good" here means "close to BERTimbau") when utilized for downstream applications.
148
+
149
+ | Models | IMDB | FaQuAD-NLI | HateBr | Assin2 | AgNews | Average |
150
+ |-----------------|-----------|------------|-----------|-----------|-----------|---------|
151
+ | BERTimbau-large | **93.58** | 92.26 | 91.57 | **88.97** | 94.11 | 92.10 |
152
+ | BERTimbau-small | 92.22 | **93.07** | 91.28 | 87.45 | 94.19 | 91.64 |
153
+ | **TTL-460m** | 91.64 | 91.18 | **92.28** | 86.43 | **94.42** | 91.19 |
154
+ | **TTL-160m** | 91.14 | 90.00 | 90.71 | 85.78 | 94.05 | 90.34 |
155
+
156
+ All the shown results are the higher accuracy scores achieved on the respective task test sets after fine-tuning the models on the training sets. All fine-tuning runs used the same hyperparameters, and the code implementation can be found in the [model cards](https://huggingface.co/nicholasKluge/TeenyTinyLlama-460m-HateBR) of our fine-tuned models.
157
 
158
  ## Cite as 🤗
159
 
160
  ```latex
161
 
162
+ @misc{correa24ttllama,
163
+ title = {TeenyTinyLlama: a pair of open-source tiny language models trained in Brazilian Portuguese},
164
+ author = {Corr{\^e}a, Nicholas Kluge and Falk, Sophia and Fatimah, Shiza and Sen, Aniket and De Oliveira, Nythamar},
165
+ journal={arXiv},
166
+ year = {2024},
 
 
 
167
  }
168
 
169
  ```