Text Generation
Transformers
Safetensors
English
falcon_mamba
Eval Results
Inference Endpoints
Gkunsch commited on
Commit
8a4e871
·
verified ·
1 Parent(s): 34e00ff

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -138,7 +138,8 @@ Similar to the others [Falcon](https://huggingface.co/tiiuae/falcon-11B) suite m
138
  Note that at inference the context-length is not relevant as the Mamba architecture has no limit on long range dependency.
139
  At the last training stage, small portion of high-quality curated data was used to further enhance performance.
140
 
141
- Overall, the data sources included RefinedWeb-English, Refined-Multilingual (latin languages), high quality technical data, code data, and conversational data extracted from public sources.
 
142
 
143
  The data was tokenized with the Falcon-[7B](https://huggingface.co/tiiuae/falcon-7B)/[11B](https://huggingface.co/tiiuae/falcon-11B) tokenizer.
144
 
 
138
  Note that at inference the context-length is not relevant as the Mamba architecture has no limit on long range dependency.
139
  At the last training stage, small portion of high-quality curated data was used to further enhance performance.
140
 
141
+ Overall, the data sources included RefinedWeb-English, high quality technical data, code data and conversational data extracted from public sources.
142
+ In particular, we used samples coming from [Fineweb-edu](https://huggingface.co/datasets/HuggingFaceFW/fineweb-edu).
143
 
144
  The data was tokenized with the Falcon-[7B](https://huggingface.co/tiiuae/falcon-7B)/[11B](https://huggingface.co/tiiuae/falcon-11B) tokenizer.
145