IsmailH commited on
Commit
0702252
1 Parent(s): bfb35c4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -5
README.md CHANGED
@@ -43,11 +43,6 @@ inference:
43
  <!-- Quantized versions in GGUF format can be found in [TheBloke/Claire-7B-0.1-GGUF](https://huggingface.co/TheBloke/Claire-7B-0.1-GGUF). -->
44
 
45
  Claire-7B-EN-0.1 is a pretrained language model designed to be attuned to the dynamics of linguistic interactions in dialogue. Without further training, its expected use is to generate continuations of dialogues. Its main purpose is to serve as a base model for fine-tuning on dialogue generation (e.g., chat) and dialogue understanding (e.g., meeting summarization) tasks. Please note that due to its training, the model is prone to generate dialogues with disfluencies and other constructions common to spoken language.
46
- Claire-7B-EN-0.1 is finetuned only on English dialogue data, but the following variants are available to evaluate the impact of language mixture on dialogue understanding.
47
- * [Claire-7B-FR-EN-25-75](OpenLLM-France/Claire-7B-FR-EN-25-75-0.1), with 25/75 French-English data split.
48
- * [Claire-7B-FR-EN-50-50](OpenLLM-France/Claire-7B-FR-EN-50-50-0.1), with 50/50 French-English data split.
49
- * [Claire-7B-FR-EN-75-25](OpenLLM-France/Claire-7B-FR-EN-75-25-0.1), with 75/25 French-English data split.
50
- * [Claire-FR-7B](OpenLLM-France/Claire-7B-0.1), with only French data.
51
 
52
 
53
  * [Typical usage](#typical-usage)
@@ -56,6 +51,7 @@ Claire-7B-EN-0.1 is finetuned only on English dialogue data, but the following v
56
  * [Training Data](#training-data)
57
  * [Training Procedure](#training-procedure)
58
  <!-- * [Evaluation](#evaluation) -->
 
59
  * [License](#license)
60
  * [Acknowledgements](#acknowledgements)
61
  * [Contact](#contact)
@@ -230,6 +226,15 @@ and "Claire-Mistral", for [Claire-Mistral-7B-0.1](https://huggingface.co/OpenLLM
230
  Please note that the model can generate disfluencies and humorous responses as a result of its training on spoken and theatrical text.
231
  -->
232
 
 
 
 
 
 
 
 
 
 
233
  ## License
234
 
235
  Given that some of the corpora used for training are only available under CC-BY-NC-SA licenses,
 
43
  <!-- Quantized versions in GGUF format can be found in [TheBloke/Claire-7B-0.1-GGUF](https://huggingface.co/TheBloke/Claire-7B-0.1-GGUF). -->
44
 
45
  Claire-7B-EN-0.1 is a pretrained language model designed to be attuned to the dynamics of linguistic interactions in dialogue. Without further training, its expected use is to generate continuations of dialogues. Its main purpose is to serve as a base model for fine-tuning on dialogue generation (e.g., chat) and dialogue understanding (e.g., meeting summarization) tasks. Please note that due to its training, the model is prone to generate dialogues with disfluencies and other constructions common to spoken language.
 
 
 
 
 
46
 
47
 
48
  * [Typical usage](#typical-usage)
 
51
  * [Training Data](#training-data)
52
  * [Training Procedure](#training-procedure)
53
  <!-- * [Evaluation](#evaluation) -->
54
+ * [Variants](#variants)
55
  * [License](#license)
56
  * [Acknowledgements](#acknowledgements)
57
  * [Contact](#contact)
 
226
  Please note that the model can generate disfluencies and humorous responses as a result of its training on spoken and theatrical text.
227
  -->
228
 
229
+ ## Variants
230
+
231
+ Claire-7B-EN-0.1 is finetuned only on English dialogue data, but the following variants are available to evaluate the impact of language mixture on dialogue understanding.
232
+ * [Claire-7B-FR-EN-25-75](OpenLLM-France/Claire-7B-FR-EN-25-75-0.1), with 25/75 French-English data split.
233
+ * [Claire-7B-FR-EN-50-50](OpenLLM-France/Claire-7B-FR-EN-50-50-0.1), with 50/50 French-English data split.
234
+ * [Claire-7B-FR-EN-75-25](OpenLLM-France/Claire-7B-FR-EN-75-25-0.1), with 75/25 French-English data split.
235
+ * [Claire-FR-7B](OpenLLM-France/Claire-7B-0.1), with only French data.
236
+
237
+
238
  ## License
239
 
240
  Given that some of the corpora used for training are only available under CC-BY-NC-SA licenses,