juliehunter commited on
Commit
6615814
·
verified ·
1 Parent(s): 204e188

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -9
README.md CHANGED
@@ -5,8 +5,6 @@ language:
5
  - fr
6
  - en
7
  tags:
8
- - pretrained
9
- - llama-3
10
  - openllm-france
11
  datasets:
12
  - cmh/alpaca_data_cleaned_fr_52k
@@ -20,11 +18,11 @@ datasets:
20
  base_model:
21
  - OpenLLM-France/Lucie-7B
22
  widget:
23
- - text: |-
24
- Quelle est la capitale de l'Espagne ? Madrid.
25
- Quelle est la capitale de la France ?
26
- example_title: Capital cities in French
27
- group: 1-shot Question Answering
28
  training_progress:
29
  context_length: 32000
30
  ---
@@ -57,7 +55,7 @@ Note that this instruction training is light and is meant to allow Lucie to prod
57
 
58
  Due to its size, Lucie-7B is limited in the information that it can memorize; its ability to produce correct answers could be improved by implementing the model in a retrieval augmented generation pipeline.
59
 
60
- While Lucie-7B-Instruct is trained on sequences of 4096 tokens, its base model, Lucie-7B has a context size of 32K tokens. Based on Needle-in-a-haystack evaluations, Lucie-7B-Instruct maintains the capacity of the base model to handle 32K-size context windows.
61
 
62
 
63
  ## Training details
@@ -232,4 +230,4 @@ Finally, we thank the entire OpenLLM-France community, whose members have helped
232
 
233
  ## Contact
234
 
235
- contact@openllm-france.fr
 
5
  - fr
6
  - en
7
  tags:
 
 
8
  - openllm-france
9
  datasets:
10
  - cmh/alpaca_data_cleaned_fr_52k
 
18
  base_model:
19
  - OpenLLM-France/Lucie-7B
20
  widget:
21
+ - text: |-
22
+ Quelle est la capitale de l'Espagne ? Madrid.
23
+ Quelle est la capitale de la France ?
24
+ example_title: Capital cities in French
25
+ group: 1-shot Question Answering
26
  training_progress:
27
  context_length: 32000
28
  ---
 
55
 
56
  Due to its size, Lucie-7B is limited in the information that it can memorize; its ability to produce correct answers could be improved by implementing the model in a retrieval augmented generation pipeline.
57
 
58
+ While Lucie-7B-Instruct is trained on sequences of 4096 tokens, its base model, Lucie-7B has a context size of 32K tokens. Based on Needle-in-a-haystack evaluations, Lucie-7B-Instruct-v1.1 has a context window size of 22K tokens. This window could be increasd by fine-tuning on longer data samples.
59
 
60
 
61
  ## Training details
 
230
 
231
  ## Contact
232
 
233
+ contact@openllm-france.fr