Update README.md
Browse files
README.md
CHANGED
@@ -43,11 +43,6 @@ inference:
|
|
43 |
<!-- Quantized versions in GGUF format can be found in [TheBloke/Claire-7B-0.1-GGUF](https://huggingface.co/TheBloke/Claire-7B-0.1-GGUF). -->
|
44 |
|
45 |
Claire-7B-EN-0.1 is a pretrained language model designed to be attuned to the dynamics of linguistic interactions in dialogue. Without further training, its expected use is to generate continuations of dialogues. Its main purpose is to serve as a base model for fine-tuning on dialogue generation (e.g., chat) and dialogue understanding (e.g., meeting summarization) tasks. Please note that due to its training, the model is prone to generate dialogues with disfluencies and other constructions common to spoken language.
|
46 |
-
Claire-7B-EN-0.1 is finetuned only on English dialogue data, but the following variants are available to evaluate the impact of language mixture on dialogue understanding.
|
47 |
-
* [Claire-7B-FR-EN-25-75](OpenLLM-France/Claire-7B-FR-EN-25-75-0.1), with 25/75 French-English data split.
|
48 |
-
* [Claire-7B-FR-EN-50-50](OpenLLM-France/Claire-7B-FR-EN-50-50-0.1), with 50/50 French-English data split.
|
49 |
-
* [Claire-7B-FR-EN-75-25](OpenLLM-France/Claire-7B-FR-EN-75-25-0.1), with 75/25 French-English data split.
|
50 |
-
* [Claire-FR-7B](OpenLLM-France/Claire-7B-0.1), with only French data.
|
51 |
|
52 |
|
53 |
* [Typical usage](#typical-usage)
|
@@ -56,6 +51,7 @@ Claire-7B-EN-0.1 is finetuned only on English dialogue data, but the following v
|
|
56 |
* [Training Data](#training-data)
|
57 |
* [Training Procedure](#training-procedure)
|
58 |
<!-- * [Evaluation](#evaluation) -->
|
|
|
59 |
* [License](#license)
|
60 |
* [Acknowledgements](#acknowledgements)
|
61 |
* [Contact](#contact)
|
@@ -230,6 +226,15 @@ and "Claire-Mistral", for [Claire-Mistral-7B-0.1](https://huggingface.co/OpenLLM
|
|
230 |
Please note that the model can generate disfluencies and humorous responses as a result of its training on spoken and theatrical text.
|
231 |
-->
|
232 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
233 |
## License
|
234 |
|
235 |
Given that some of the corpora used for training are only available under CC-BY-NC-SA licenses,
|
|
|
43 |
<!-- Quantized versions in GGUF format can be found in [TheBloke/Claire-7B-0.1-GGUF](https://huggingface.co/TheBloke/Claire-7B-0.1-GGUF). -->
|
44 |
|
45 |
Claire-7B-EN-0.1 is a pretrained language model designed to be attuned to the dynamics of linguistic interactions in dialogue. Without further training, its expected use is to generate continuations of dialogues. Its main purpose is to serve as a base model for fine-tuning on dialogue generation (e.g., chat) and dialogue understanding (e.g., meeting summarization) tasks. Please note that due to its training, the model is prone to generate dialogues with disfluencies and other constructions common to spoken language.
|
|
|
|
|
|
|
|
|
|
|
46 |
|
47 |
|
48 |
* [Typical usage](#typical-usage)
|
|
|
51 |
* [Training Data](#training-data)
|
52 |
* [Training Procedure](#training-procedure)
|
53 |
<!-- * [Evaluation](#evaluation) -->
|
54 |
+
* [Variants](#variants)
|
55 |
* [License](#license)
|
56 |
* [Acknowledgements](#acknowledgements)
|
57 |
* [Contact](#contact)
|
|
|
226 |
Please note that the model can generate disfluencies and humorous responses as a result of its training on spoken and theatrical text.
|
227 |
-->
|
228 |
|
229 |
+
## Variants
|
230 |
+
|
231 |
+
Claire-7B-EN-0.1 is finetuned only on English dialogue data, but the following variants are available to evaluate the impact of language mixture on dialogue understanding.
|
232 |
+
* [Claire-7B-FR-EN-25-75](OpenLLM-France/Claire-7B-FR-EN-25-75-0.1), with 25/75 French-English data split.
|
233 |
+
* [Claire-7B-FR-EN-50-50](OpenLLM-France/Claire-7B-FR-EN-50-50-0.1), with 50/50 French-English data split.
|
234 |
+
* [Claire-7B-FR-EN-75-25](OpenLLM-France/Claire-7B-FR-EN-75-25-0.1), with 75/25 French-English data split.
|
235 |
+
* [Claire-FR-7B](OpenLLM-France/Claire-7B-0.1), with only French data.
|
236 |
+
|
237 |
+
|
238 |
## License
|
239 |
|
240 |
Given that some of the corpora used for training are only available under CC-BY-NC-SA licenses,
|