m-polignano-uniba
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -24,9 +24,9 @@ license: llama3
|
|
24 |
<hr>
|
25 |
<!--<img src="https://i.ibb.co/6mHSRm3/llamantino53.jpg" width="200"/>-->
|
26 |
|
27 |
-
|
28 |
The model is an instruction-tuned version of [**Meta-Llama-3-8b-instruct**](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) (a fine-tuned **LLaMA 3 model**).
|
29 |
-
This model version aims to be the **Multilingual
|
30 |
|
31 |
|
32 |
The ๐**ANITA project**๐ *(**A**dvanced **N**atural-based interaction for the **ITA**lian language)*
|
@@ -46,6 +46,7 @@ wants to provide Italian NLP researchers with an improved model the for Italian
|
|
46 |
- **Model developers**: Ph.D. Marco Polignano - University of Bari Aldo Moro, Italy - SWAP Research Group
|
47 |
- **Variations**: The model release has been **supervised fine-tuning (SFT)** using **QLoRA** 4bit, on two instruction-based datasets. **DPO** approach over the *jondurbin/truthy-dpo-v0.1* dataset is used to align with human preferences for helpfulness and safety.
|
48 |
- **Input**: Models input text only.
|
|
|
49 |
- **Output**: Models generate text and code only.
|
50 |
- **Model Architecture**: *Llama 3 architecture*.
|
51 |
- **Context length**: 8K, 8192.
|
|
|
24 |
<hr>
|
25 |
<!--<img src="https://i.ibb.co/6mHSRm3/llamantino53.jpg" width="200"/>-->
|
26 |
|
27 |
+
<p style="text-align:justify;">**LLaMAntino-3-ANITA-8B-Instr-DPO-ITA** is a model of the [**LLaMAntino**](https://huggingface.co/swap-uniba) - *Large Language Models family*.
|
28 |
The model is an instruction-tuned version of [**Meta-Llama-3-8b-instruct**](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) (a fine-tuned **LLaMA 3 model**).
|
29 |
+
This model version aims to be the a **Multilingual Model** ๐ -- EN ๐บ๐ธ + ITA๐ฎ๐น -- to further fine-tune for the Specific Italian Task</p>
|
30 |
|
31 |
|
32 |
The ๐**ANITA project**๐ *(**A**dvanced **N**atural-based interaction for the **ITA**lian language)*
|
|
|
46 |
- **Model developers**: Ph.D. Marco Polignano - University of Bari Aldo Moro, Italy - SWAP Research Group
|
47 |
- **Variations**: The model release has been **supervised fine-tuning (SFT)** using **QLoRA** 4bit, on two instruction-based datasets. **DPO** approach over the *jondurbin/truthy-dpo-v0.1* dataset is used to align with human preferences for helpfulness and safety.
|
48 |
- **Input**: Models input text only.
|
49 |
+
- **Language**: Multilingual๐ + Italian ๐ฎ๐น
|
50 |
- **Output**: Models generate text and code only.
|
51 |
- **Model Architecture**: *Llama 3 architecture*.
|
52 |
- **Context length**: 8K, 8192.
|