manu commited on
Commit
b5dc3f9
1 Parent(s): 7be6e70

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +77 -0
README.md ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - cerebras/SlimPajama-627B
5
+ - uonlp/CulturaX
6
+ - pg19
7
+ - bigcode/starcoderdata
8
+ - croissantllm/croissant_dataset
9
+ language:
10
+ - fr
11
+ - en
12
+ pipeline_tag: text-generation
13
+ tags:
14
+ - legal
15
+ - code
16
+ - text-generation-inference
17
+ - art
18
+ ---
19
+
20
+ # CroissantLLM - All smaller checkpoints
21
+
22
+ These models are part of the CroissantLLM initiative, and correspond to the checkpoints after 100B tokens for smaller model sizes.
23
+ These are the models used for scaling laws.
24
+
25
+
26
+ To play with the final model, we recommend using the Chat version: https://huggingface.co/croissantllm/CroissantLLMChat-v0.1.
27
+
28
+
29
+ https://arxiv.org/abs/2402.00786
30
+
31
+
32
+
33
+ ## Abstract
34
+ We introduce CroissantLLM, a 1.3B language model pretrained on a set of 3T English and French tokens, to bring to the research and industrial community a high-performance, fully open-sourced bilingual model that runs swiftly on consumer-grade local hardware.
35
+ To that end, we pioneer the approach of training an intrinsically bilingual model with a 1:1 English-to-French pretraining data ratio, a custom tokenizer, and bilingual finetuning datasets. We release the training dataset, notably containing a French split with manually curated, high-quality, and varied data sources.
36
+ To assess performance outside of English, we craft a novel benchmark, FrenchBench, consisting of an array of classification and generation tasks, covering various orthogonal aspects of model performance in the French Language. Additionally, rooted in transparency and to foster further Large Language Model research, we release codebases, and dozens of checkpoints across various model sizes, training data distributions, and training steps, as well as fine-tuned Chat models, and strong translation models. We evaluate our model through the FMTI framework, and validate 81% of the transparency criteria, far beyond the scores of even most open initiatives.
37
+ This work enriches the NLP landscape, breaking away from previous English-centric work in order to strengthen our understanding of multilinguality in language models.
38
+
39
+ ## Citation
40
+
41
+ Our work can be cited as:
42
+
43
+ ```bash
44
+ @misc{faysse2024croissantllm,
45
+ title={CroissantLLM: A Truly Bilingual French-English Language Model},
46
+ author={Manuel Faysse and Patrick Fernandes and Nuno M. Guerreiro and António Loison and Duarte M. Alves and Caio Corro and Nicolas Boizard and João Alves and Ricardo Rei and Pedro H. Martins and Antoni Bigata Casademunt and François Yvon and André F. T. Martins and Gautier Viaud and Céline Hudelot and Pierre Colombo},
47
+ year={2024},
48
+ eprint={2402.00786},
49
+ archivePrefix={arXiv},
50
+ primaryClass={cs.CL}
51
+ }
52
+ ```
53
+
54
+ ## Usage
55
+
56
+ This model is a base model, that is, it is not finetuned for Chat function and works best with few-shot prompting strategies.
57
+
58
+
59
+ ```python
60
+
61
+ import torch
62
+ from transformers import AutoModelForCausalLM, AutoTokenizer
63
+
64
+
65
+ model_name = "croissantllm/CroissantLLMBase"
66
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
67
+ model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16, device_map="auto")
68
+
69
+ inputs = tokenizer("I am so tired I could sleep right now. -> Je suis si fatigué que je pourrais m'endormir maintenant.\nHe is heading to the market. -> Il va au marché.\nWe are running on the beach. ->", return_tensors="pt").to(model.device)
70
+ tokens = model.generate(**inputs, max_length=100, do_sample=True, top_p=0.95, top_k=60, temperature=0.3)
71
+ print(tokenizer.decode(tokens[0]))
72
+
73
+ # remove bos token
74
+ inputs = tokenizer("Capitales: France -> Paris, Italie -> Rome, Allemagne -> Berlin, Espagne ->", return_tensors="pt", add_special_tokens=True).to(model.device)
75
+ tokens = model.generate(**inputs, max_length=100, do_sample=True, top_p=0.95, top_k=60)
76
+ print(tokenizer.decode(tokens[0]))
77
+ ```