jpacifico commited on
Commit
d894e6c
·
verified ·
1 Parent(s): 53e87bf

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -3
README.md CHANGED
@@ -14,13 +14,12 @@ datasets:
14
  # Description model
15
 
16
  Chocolatine-Admin-3B version specialized in French administrative language, supervised fine-tuning of [jpacifico/Chocolatine-3B-Instruct-DPO-v1.2](https://huggingface.co/jpacifico/Chocolatine-3B-Instruct-DPO-v1.2) based on [microsoft/Phi-3.5-mini-instruct](https://huggingface.co/microsoft/Phi-3.5-mini-instruct)
17
- based on the official [lexicon](https://www.modernisation.gouv.fr/outils-et-formations/lexique-administratif) published by the French Ministère de la Fonction Publique et de la Réforme de l'Etat.
18
  Developed in collaboration with Microsoft.
19
 
20
 
21
  # Data & Training
22
 
23
- The [dataset](jpacifico/merged-admin-def-dataset-16k) gathers 2362 administrative terms constituting the basis of the simulation of prompt-answer pairs.
24
  The GPT-4o model deployed on Azure OpenAI was used to carry out the building of the dataset in several phases:
25
 
26
  - Extraction of the lexicon pages (previously converted into jpg format)
@@ -28,7 +27,7 @@ The GPT-4o model deployed on Azure OpenAI was used to carry out the building of
28
  - Generation of questions from the terms and definitions
29
  - Generation of answers in three successive rounds taking into account the previous generations to ensure variety.
30
 
31
- For this version the Fine Tuning (SFT) was performed on 11 epochs with an A100 GPU instance on Azure Machine Learning.
32
 
33
  # Usage
34
 
 
14
  # Description model
15
 
16
  Chocolatine-Admin-3B version specialized in French administrative language, supervised fine-tuning of [jpacifico/Chocolatine-3B-Instruct-DPO-v1.2](https://huggingface.co/jpacifico/Chocolatine-3B-Instruct-DPO-v1.2) based on [microsoft/Phi-3.5-mini-instruct](https://huggingface.co/microsoft/Phi-3.5-mini-instruct)
 
17
  Developed in collaboration with Microsoft.
18
 
19
 
20
  # Data & Training
21
 
22
+ The [dataset](jpacifico/merged-admin-def-dataset-16k) based on the official [lexicon](https://www.modernisation.gouv.fr/outils-et-formations/lexique-administratif) published by the French DITP, gathers 2362 administrative terms constituting the basis of the simulation of prompt-answer pairs.
23
  The GPT-4o model deployed on Azure OpenAI was used to carry out the building of the dataset in several phases:
24
 
25
  - Extraction of the lexicon pages (previously converted into jpg format)
 
27
  - Generation of questions from the terms and definitions
28
  - Generation of answers in three successive rounds taking into account the previous generations to ensure variety.
29
 
30
+ For this 0.3b version, the Fine Tuning (SFT) was performed on 11 epochs with an A100 GPU instance on Azure Machine Learning.
31
 
32
  # Usage
33