Update README.md
Browse files
README.md
CHANGED
@@ -14,13 +14,12 @@ datasets:
|
|
14 |
# Description model
|
15 |
|
16 |
Chocolatine-Admin-3B version specialized in French administrative language, supervised fine-tuning of [jpacifico/Chocolatine-3B-Instruct-DPO-v1.2](https://huggingface.co/jpacifico/Chocolatine-3B-Instruct-DPO-v1.2) based on [microsoft/Phi-3.5-mini-instruct](https://huggingface.co/microsoft/Phi-3.5-mini-instruct)
|
17 |
-
based on the official [lexicon](https://www.modernisation.gouv.fr/outils-et-formations/lexique-administratif) published by the French Ministère de la Fonction Publique et de la Réforme de l'Etat.
|
18 |
Developed in collaboration with Microsoft.
|
19 |
|
20 |
|
21 |
# Data & Training
|
22 |
|
23 |
-
The [dataset](jpacifico/merged-admin-def-dataset-16k) gathers 2362 administrative terms constituting the basis of the simulation of prompt-answer pairs.
|
24 |
The GPT-4o model deployed on Azure OpenAI was used to carry out the building of the dataset in several phases:
|
25 |
|
26 |
- Extraction of the lexicon pages (previously converted into jpg format)
|
@@ -28,7 +27,7 @@ The GPT-4o model deployed on Azure OpenAI was used to carry out the building of
|
|
28 |
- Generation of questions from the terms and definitions
|
29 |
- Generation of answers in three successive rounds taking into account the previous generations to ensure variety.
|
30 |
|
31 |
-
For this version the Fine Tuning (SFT) was performed on 11 epochs with an A100 GPU instance on Azure Machine Learning.
|
32 |
|
33 |
# Usage
|
34 |
|
|
|
14 |
# Description model
|
15 |
|
16 |
Chocolatine-Admin-3B version specialized in French administrative language, supervised fine-tuning of [jpacifico/Chocolatine-3B-Instruct-DPO-v1.2](https://huggingface.co/jpacifico/Chocolatine-3B-Instruct-DPO-v1.2) based on [microsoft/Phi-3.5-mini-instruct](https://huggingface.co/microsoft/Phi-3.5-mini-instruct)
|
|
|
17 |
Developed in collaboration with Microsoft.
|
18 |
|
19 |
|
20 |
# Data & Training
|
21 |
|
22 |
+
The [dataset](jpacifico/merged-admin-def-dataset-16k) based on the official [lexicon](https://www.modernisation.gouv.fr/outils-et-formations/lexique-administratif) published by the French DITP, gathers 2362 administrative terms constituting the basis of the simulation of prompt-answer pairs.
|
23 |
The GPT-4o model deployed on Azure OpenAI was used to carry out the building of the dataset in several phases:
|
24 |
|
25 |
- Extraction of the lexicon pages (previously converted into jpg format)
|
|
|
27 |
- Generation of questions from the terms and definitions
|
28 |
- Generation of answers in three successive rounds taking into account the previous generations to ensure variety.
|
29 |
|
30 |
+
For this 0.3b version, the Fine Tuning (SFT) was performed on 11 epochs with an A100 GPU instance on Azure Machine Learning.
|
31 |
|
32 |
# Usage
|
33 |
|