robinsmits commited on
Commit
12e189e
1 Parent(s): 35ce203

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +44 -7
README.md CHANGED
@@ -1,36 +1,61 @@
1
  ---
2
- license: apache-2.0
 
 
 
3
  base_model: DAMO-NLP-MT/polylm-1.7b
4
  tags:
5
  - generated_from_trainer
 
 
 
 
6
  model-index:
7
  - name: polylm_1.7b_ft_alpaca_clean_dutch
8
  results: []
 
 
 
 
9
  ---
10
 
11
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
- should probably proofread and complete it, then remove this comment. -->
13
-
14
  # polylm_1.7b_ft_alpaca_clean_dutch
15
 
16
- This model is a fine-tuned version of [DAMO-NLP-MT/polylm-1.7b](https://huggingface.co/DAMO-NLP-MT/polylm-1.7b) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
  - Loss: 1.8483
19
 
 
 
 
 
20
  ## Model description
21
 
22
  More information needed
23
 
24
  ## Intended uses & limitations
25
 
26
- More information needed
 
 
 
 
 
27
 
28
  ## Training and evaluation data
29
 
30
- More information needed
 
 
 
 
31
 
32
  ## Training procedure
33
 
 
 
 
 
34
  ### Training hyperparameters
35
 
36
  The following hyperparameters were used during training:
@@ -45,6 +70,17 @@ The following hyperparameters were used during training:
45
  - lr_scheduler_warmup_steps: 64
46
  - num_epochs: 2
47
 
 
 
 
 
 
 
 
 
 
 
 
48
  ### Training results
49
 
50
  | Training Loss | Epoch | Step | Validation Loss |
@@ -69,3 +105,4 @@ The following hyperparameters were used during training:
69
  - Pytorch 2.0.1+cu118
70
  - Datasets 2.13.1
71
  - Tokenizers 0.13.3
 
 
1
  ---
2
+ license: cc-by-nc-4.0
3
+ inference: false
4
+ datasets:
5
+ - BramVanroy/alpaca-cleaned-dutch
6
  base_model: DAMO-NLP-MT/polylm-1.7b
7
  tags:
8
  - generated_from_trainer
9
+ - alpaca
10
+ - Transformers
11
+ - PolyLM
12
+ - text-generation-inference
13
  model-index:
14
  - name: polylm_1.7b_ft_alpaca_clean_dutch
15
  results: []
16
+ language:
17
+ - nl
18
+ library_name: peft
19
+ pipeline_tag: text-generation
20
  ---
21
 
 
 
 
22
  # polylm_1.7b_ft_alpaca_clean_dutch
23
 
24
+ This adapter model is a fine-tuned version of [DAMO-NLP-MT/polylm-1.7b](https://huggingface.co/DAMO-NLP-MT/polylm-1.7b).
25
  It achieves the following results on the evaluation set:
26
  - Loss: 1.8483
27
 
28
+ Finetuning was performed on the Dutch [BramVanroy/alpaca-cleaned-dutch](https://www.huggingface.co/datasets/BramVanroy/alpaca-cleaned-dutch) dataset which contains 52K of records with instruction following-data translated from English to Dutch.
29
+
30
+ See [DAMO-NLP-MT/polylm-1.7b](https://huggingface.co/DAMO-NLP-MT/polylm-1.7b) for all information about the base model.
31
+
32
  ## Model description
33
 
34
  More information needed
35
 
36
  ## Intended uses & limitations
37
 
38
+ The PolyLM-1.7B model was trained on 18 languages. The primary focus was to create a multi-lingual Open LLM.
39
+ Dutch was one of those 18 languages. For training the model a diverse combination of multi-lingual datasets was used.
40
+
41
+ The generated output and performance of this model for the Dutch language is very likely not always comparable to the various Open-Llama models that have been finetuned on English Alpaca datasets.
42
+
43
+ The primary intention of this finetuned model is to explore and research the use of the Dutch language in combination with an Open LLM model.
44
 
45
  ## Training and evaluation data
46
 
47
+ This model was trained on the [BramVanroy/alpaca-cleaned-dutch](https://www.huggingface.co/datasets/BramVanroy/alpaca-cleaned-dutch) dataset.
48
+
49
+ The dataset is the Dutch translation of the English Alpaca Cleaned instruction dataset.
50
+
51
+ Based on the dataset license only Non-Commercial use is allowed. Commercial use is strictly forbidden.
52
 
53
  ## Training procedure
54
 
55
+ This model was finetuned with a QLoRA setup on a Google Colab A100 GPU in about 1.5 hours.
56
+
57
+ The notebook used for training can be found here: [Training Notebook](https://github.com/RobinSmits/Dutch-LLMs/blob/main/PolyLM_1_7B_Alpaca_Clean_Dutch_Qlora.ipynb)
58
+
59
  ### Training hyperparameters
60
 
61
  The following hyperparameters were used during training:
 
70
  - lr_scheduler_warmup_steps: 64
71
  - num_epochs: 2
72
 
73
+ The following bitsandbytes quantization config was used during training:
74
+ - load_in_8bit: False
75
+ - load_in_4bit: True
76
+ - llm_int8_threshold: 6.0
77
+ - llm_int8_skip_modules: None
78
+ - llm_int8_enable_fp32_cpu_offload: False
79
+ - llm_int8_has_fp16_weight: False
80
+ - bnb_4bit_quant_type: nf4
81
+ - bnb_4bit_use_double_quant: True
82
+ - bnb_4bit_compute_dtype: bfloat16
83
+
84
  ### Training results
85
 
86
  | Training Loss | Epoch | Step | Validation Loss |
 
105
  - Pytorch 2.0.1+cu118
106
  - Datasets 2.13.1
107
  - Tokenizers 0.13.3
108
+ - PEFT 0.4.0