HiTZ
/

Text Generation
Transformers
PyTorch
Basque
English
llama
text-generation-inference
Inference Endpoints
latxa-7b-v1 / README.md
OSainz's picture
Update README.md
dbac5c2 verified
---
license: llama2
datasets:
- HiTZ/euscrawl
language:
- eu
- en
metrics:
- accuracy
- f1
- perplexity
pipeline_tag: text-generation
---
# **Model Card for Latxa 7b**
<p align="center">
<img src="https://github.com/hitz-zentroa/latxa/blob/b9aa705f60ee2cc03c9ed62fda82a685abb31b07/assets/latxa_round.png?raw=true" style="height: 350px;">
</p>
<span style="color: red; font-weight: bold">IMPORTANT:</span> This model is outdated and made available publicly for reproducibility purposes only. Please utilize the most recent version found in [our HuggingFace collection](https://huggingface.co/collections/HiTZ/latxa-65a697e6838b3acc53677304).
Latxa is a collection of foundation models specifically tuned for Basque. Based on Meta’s LLaMA 2 model family, these models were further trained with Euscrawl, a highly curated Basque corpora ([Artetxe et al., 2022](https://aclanthology.org/2022.emnlp-main.499/)). Ranging from 7 billion to 70 billion parameters, these models are currently the biggest and best-performing LLMs built for Basque. This is the 7b repository, links to other models can be found in the [Latxa Collection](https://huggingface.co/collections/HiTZ/latxa-65a697e6838b3acc53677304).
Read more about Latxa in our [website](https://www.hitz.eus/en/node/340) or in [LinkedIn](https://www.linkedin.com/pulse/presenting-latxa-largest-language-model-built-basque-hitz-zentroa-63qdf)!
# **Model Details**
## **Model Description**
Latxa is a family of Large Language Models (LLM) based on Meta’s [LLaMA models](https://huggingface.co/meta-llama). Current LLMs exhibit incredible performance for high-resource languages such as English, but, in the case of Basque and other low-resource languages, their performance is close to a random guesser. These limitations widen the gap between high- and low-resource languages when it comes to digital development. We present Latxa to overcome these limitations and promote the development of LLM-based technology and research for the Basque language. Latxa models follow the same architecture as their original counterparts and were further trained in Euscrawl v1 ([Artetxe et al., 2022](https://aclanthology.org/2022.emnlp-main.499/)), a high-quality Basque corpora.
The models are released in three sizes: 7B, 13B and 70B.
* **Developed by:** HiTZ Research Center & IXA Research group (University of the Basque Country UPV/EHU)
* **Model type:** Language model
* **Language(s) (NLP):** en, eu
* **License:** llama2
* **Parent Model:** meta-llama/Llama-2-7b
* **Contact:** hitz@ehu.eus
## **Getting started**
Use the code below to get started with the model.
```python
from transformers import pipeline
pipe = pipeline("text-generation", model=”HiTZ/latxa-7b-v1”)
text = "Euskara adimen artifizialera iritsi da!"
pipe(text, max_new_tokens=50, num_beams=5)
>> [
{
'generated_text': 'Euskara adimen artifizialera iritsi da!\nEuskararen eta adimen artifizialaren arteko harremana aspaldikoa da,'
' baina azken urteotan aurrerapauso handiak eman dira arlo horretan'
}
]
```
# **Uses**
Latxa models are intended to be used with Basque data; for any other language the performance is not guaranteed. Same as the original, Latxa inherits the [LLaMA-2 License](https://ai.meta.com/llama/license/) which allows for commercial and research use.
## **Direct Use**
Latxa family models are pre-trained LLMs without any task-specific or instruction fine-tuning. That is, the model can either be prompted to perform a specific task or further fine-tuned for specific use cases.
## **Out-of-Scope Use**
The model was not fine-tuned to follow instructions or to work as a chat assistant, therefore, this kind of usage is not tested nor recommended.
# **Bias, Risks, and Limitations**
In an effort to alleviate the potentially disturbing or harmful content, Latxa has been trained on carefully selected and processed data which comes mainly from local media, national/regional newspapers, encyclopedias and blogs (see Euscrawl below). Still, the model is based on LLaMA models and can potentially carry the same bias, risk and limitations.
Please see the LLaMA’s _Ethical Considerations and Limitations _for further information.
# **Training Details**
## **Training Data**
The models were trained on EusCrawl v1, a high-quality corpus for Basque comprising 1.72M documents, 288M words, totalling 2.1GiB of uncompressed text. EusCrawl was built using ad-hoc scrapers to extract text from 33 Basque websites with high-quality content, resulting in cleaner text compared to general-purpose approaches.
See more details in the [EusCrawl](https://huggingface.co/datasets/HiTZ/euscrawl) dataset card.
Additionally, 100K documents of English data randomly selected from the [Pile](https://huggingface.co/datasets/EleutherAI/pile) dataset were also included to avoid catastrophic forgetting.
## **Training Procedure**
The models were trained using the GPT-Neox library on the HPC CINECA computing cluster. All the models were approximately trained with an effective batch size of 2M tokens for 1000 to 2000 steps.
<table>
<tr>
<td>Model
</td>
<td>Steps
</td>
<td>Sequence length
</td>
<td>Effective Batch size
</td>
<td>Total tokens
</td>
<td>GPU hours
</td>
</tr>
<tr>
<td>Latxa 7B
</td>
<td><p style="text-align: right">
2000</p>
</td>
<td><p style="text-align: right">
4096</p>
</td>
<td><p style="text-align: right">
2M tokens/step</p>
</td>
<td><p style="text-align: right">
4B</p>
</td>
<td><p style="text-align: right">
359.2h</p>
</td>
</tr>
<tr>
<td>Latxa 13B
</td>
<td><p style="text-align: right">
1000</p>
</td>
<td><p style="text-align: right">
4096</p>
</td>
<td><p style="text-align: right">
2M tokens/step</p>
</td>
<td><p style="text-align: right">
2B</p>
</td>
<td><p style="text-align: right">
468.8h</p>
</td>
</tr>
<tr>
<td>Latxa 70B
</td>
<td><p style="text-align: right">
1680</p>
</td>
<td><p style="text-align: right">
4096</p>
</td>
<td><p style="text-align: right">
2M tokens/step</p>
</td>
<td><p style="text-align: right">
3.4B</p>
</td>
<td><p style="text-align: right">
*6475.52h</p>
</td>
</tr>
</table>
* indicates the time for the entire training process (2000 steps), however the weights of the step 1680 are shared as it is the best checkpoint according to validation loss.
# **Evaluation**
We evaluated the models on zero-shot and few-shot settings on generative, multiple-choice and classification tasks. We used the basque partitions of each dataset.
## **Testing Data, Factors & Metrics**
### **Testing Data**
* **Belebele** ([Bandarkar et al.](https://arxiv.org/abs/2308.16884)): Belebele is a multiple-choice machine reading comprehension (MRC) dataset spanning 122 language variants. We evaluated the model in a 5-shot fashion.
* Data card: [https://huggingface.co/datasets/facebook/belebele](https://huggingface.co/datasets/facebook/belebele)
* **X-StoryCloze** ([Lin et al.](https://arxiv.org/abs/2112.10668)): XStoryCloze consists of the professionally translated version of the English StoryCloze dataset to 10 non-English languages. Story Cloze is a commonsense reasoning dataset which consists of choosing the correct ending to a four-sentence story. We evaluated the model in a 0-shot fashion.
* Data card: [https://huggingface.co/datasets/juletxara/xstory_cloze](https://huggingface.co/datasets/juletxara/xstory_cloze)
* **BasqueGLUE** ([Urbizu et al.](https://aclanthology.org/2022.lrec-1.172.pdf)): BasqueGLUE is a NLU benchmark for Basque. We evaluated the model in a 5-shot fashion on the following tasks:
* Data card:[ https://huggingface.co/datasets/orai-nlp/basqueGLUE](https://huggingface.co/datasets/orai-nlp/basqueGLUE).
* Tasks:
* **BEC2016eu**: Sentiment analysis on tweets about the 2016 Basque elections campaign.
* **VaxxStance**: Stance detection on tweets around the anti-vaccine movement.
* **BTHCv2**: Topic classification of news extracts with 12 categories.
* **EpecKorrefBin**: Correference detection task similar to WSC.
* **QNLIeu**: Q&A NLI built from the Basque Wikipedia.
* **WiCeu**: Basque Word-in-Context task.
### **Metrics**
* **Accuracy**: Belebele, X-StoryCloze, EpecKorrefBin, QNLI-eu, and, WiC-eu
* **Micro F1**: BEC2016-eu and BHTCv2
* **Macro F1**: VaxxStance (favor & against)
## **Results**
The model was evaluated using the LM Evaluation harness library from Eleuther AI.
In order to reproduce our results please follow the instructions in Latxa's [Github repository](https://github.com/hitz-zentroa/latxa?tab=readme-ov-file#evaluation).
<table>
<tr>
<td><strong>Model</strong>
</td>
<td><strong>Belebele</strong>
</td>
<td><strong>X-StoryCloze</strong>
</td>
<td><strong>BEC</strong>
</td>
<td><strong>Vaxx</strong>
</td>
<td><strong>BHTC</strong>
</td>
<td><strong>coref</strong>
</td>
<td><strong>QNLI</strong>
</td>
<td><strong>WiC</strong>
</td>
<td><strong>Average</strong>
</td>
</tr>
<tr>
<td>Random
</td>
<td>25.00
</td>
<td>50.00
</td>
<td>33.33
</td>
<td>33.33
</td>
<td>8.33
</td>
<td>50.00
</td>
<td>50.00
</td>
<td>50.00
</td>
<td>37.50
</td>
</tr>
<tr>
<td>LLaMA 2 7B
</td>
<td>26.22
</td>
<td>50.43
</td>
<td>41.63
</td>
<td>18.60
</td>
<td>20.06
</td>
<td>50.94
</td>
<td>48.32
</td>
<td>49.64
</td>
<td>38.23
</td>
</tr>
<tr>
<td>LLaMA 2 13B
</td>
<td>32.00
</td>
<td>50.63
</td>
<td>41.09
</td>
<td>18.25
</td>
<td>27.35
</td>
<td>49.23
</td>
<td>48.74
</td>
<td>49.21
</td>
<td>39.56
</td>
</tr>
<tr>
<td>LLaMA 2 70B
</td>
<td>33.56
</td>
<td>51.62
</td>
<td>47.47
</td>
<td>21.01
</td>
<td>31.01
</td>
<td>52.98
</td>
<td>51.26
</td>
<td>51.57
</td>
<td>42.56
</td>
</tr>
<tr>
<td>BLOOM 7B
</td>
<td>27.00
</td>
<td>57.18
</td>
<td>37.94
</td>
<td>20.72
</td>
<td>39.10
</td>
<td>48.21
</td>
<td>47.48
</td>
<td>47.57
</td>
<td>40.65
</td>
</tr>
<tr>
<td>XGLM 7B
</td>
<td>23.88
</td>
<td>57.71
</td>
<td>39.94
</td>
<td>21.58
</td>
<td>36.73
</td>
<td>50.94
</td>
<td>50.42
</td>
<td>49.21
</td>
<td>41.30
</td>
</tr>
<tr>
<td><strong>Latxa 7B</strong>
</td>
<td>35.67
</td>
<td>63.13
</td>
<td>55.61
</td>
<td>45.93
</td>
<td>44.44
</td>
<td>50.43
</td>
<td>55.04
</td>
<td>50.14
</td>
<td>50.05
</td>
</tr>
<tr>
<td><strong>Latxa 13B</strong>
</td>
<td>53.56
</td>
<td>65.85
</td>
<td>53.23
</td>
<td>48.66
</td>
<td><strong>53.61</strong>
</td>
<td>62.52
</td>
<td>57.14
</td>
<td>54.21
</td>
<td>56.10
</td>
</tr>
<tr>
<td><strong>Latxa 70B</strong>
</td>
<td><strong>71.78</strong>
</td>
<td><strong>67.57</strong>
</td>
<td><strong>63.52</strong>
</td>
<td><strong>48.95</strong>
</td>
<td>49.51
</td>
<td><strong>79.90</strong>
</td>
<td><strong>58.82</strong>
</td>
<td><strong>55.50</strong>
</td>
<td><strong>61.94</strong>
</td>
</tr>
</table>
# **Environmental Impact**
Carbon emissions are estimated using the[ Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in[ Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
* **Hardware Type:** HPC Cluster, 4x A100 64Gb nodes
* **Hours used:** 359.2h + 468.8h + 6475.52h = 7303.52h
* **Compute cluster:** CINECA HPC
* **Compute Region:** Italy
* **Carbon Emitted:** 673.75kg CO<sub>2</sub> eq
# **Acknowledgements**
This work has been partially supported by the Basque Government (IKER-GAITU project). The models were trained on the Leonardo supercomputer at CINECA under the EuroHPC Joint Undertaking, project EHPC-EXT-2023E01-013.