Update README.md
Browse files
README.md
CHANGED
@@ -5,8 +5,6 @@ language:
|
|
5 |
- fr
|
6 |
- en
|
7 |
tags:
|
8 |
-
- pretrained
|
9 |
-
- llama-3
|
10 |
- openllm-france
|
11 |
datasets:
|
12 |
- cmh/alpaca_data_cleaned_fr_52k
|
@@ -20,11 +18,11 @@ datasets:
|
|
20 |
base_model:
|
21 |
- OpenLLM-France/Lucie-7B
|
22 |
widget:
|
23 |
-
|
24 |
-
|
25 |
-
|
26 |
-
|
27 |
-
|
28 |
training_progress:
|
29 |
context_length: 32000
|
30 |
---
|
@@ -57,7 +55,7 @@ Note that this instruction training is light and is meant to allow Lucie to prod
|
|
57 |
|
58 |
Due to its size, Lucie-7B is limited in the information that it can memorize; its ability to produce correct answers could be improved by implementing the model in a retrieval augmented generation pipeline.
|
59 |
|
60 |
-
While Lucie-7B-Instruct is trained on sequences of 4096 tokens, its base model, Lucie-7B has a context size of 32K tokens. Based on Needle-in-a-haystack evaluations, Lucie-7B-Instruct
|
61 |
|
62 |
|
63 |
## Training details
|
@@ -232,4 +230,4 @@ Finally, we thank the entire OpenLLM-France community, whose members have helped
|
|
232 |
|
233 |
## Contact
|
234 |
|
235 |
-
contact@openllm-france.fr
|
|
|
5 |
- fr
|
6 |
- en
|
7 |
tags:
|
|
|
|
|
8 |
- openllm-france
|
9 |
datasets:
|
10 |
- cmh/alpaca_data_cleaned_fr_52k
|
|
|
18 |
base_model:
|
19 |
- OpenLLM-France/Lucie-7B
|
20 |
widget:
|
21 |
+
- text: |-
|
22 |
+
Quelle est la capitale de l'Espagne ? Madrid.
|
23 |
+
Quelle est la capitale de la France ?
|
24 |
+
example_title: Capital cities in French
|
25 |
+
group: 1-shot Question Answering
|
26 |
training_progress:
|
27 |
context_length: 32000
|
28 |
---
|
|
|
55 |
|
56 |
Due to its size, Lucie-7B is limited in the information that it can memorize; its ability to produce correct answers could be improved by implementing the model in a retrieval augmented generation pipeline.
|
57 |
|
58 |
+
While Lucie-7B-Instruct is trained on sequences of 4096 tokens, its base model, Lucie-7B has a context size of 32K tokens. Based on Needle-in-a-haystack evaluations, Lucie-7B-Instruct-v1.1 has a context window size of 22K tokens. This window could be increasd by fine-tuning on longer data samples.
|
59 |
|
60 |
|
61 |
## Training details
|
|
|
230 |
|
231 |
## Contact
|
232 |
|
233 |
+
contact@openllm-france.fr
|