appvoid commited on
Commit
8dadd90
·
verified ·
1 Parent(s): 7b8b09c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +50 -58
README.md CHANGED
@@ -1,61 +1,53 @@
 
 
1
  ---
2
- base_model:
3
- - h2oai/h2o-danube3-500m-chat
4
- - appvoid/massive
5
- library_name: transformers
6
  tags:
7
- - mergekit
8
- - merge
9
-
 
 
 
10
  ---
11
- # mix-2
12
-
13
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
14
-
15
- ## Merge Details
16
- ### Merge Method
17
-
18
- This model was merged using the SLERP merge method.
19
-
20
- ### Models Merged
21
-
22
- The following models were included in the merge:
23
- * [h2oai/h2o-danube3-500m-chat](https://huggingface.co/h2oai/h2o-danube3-500m-chat)
24
- * [appvoid/massive](https://huggingface.co/appvoid/massive)
25
-
26
- ### Configuration
27
-
28
- The following YAML configuration was used to produce this model:
29
-
30
- ```yaml
31
- slices:
32
- - sources:
33
- - model: appvoid/massive
34
- layer_range:
35
- - 0
36
- - 16
37
- - model: h2oai/h2o-danube3-500m-chat
38
- layer_range:
39
- - 0
40
- - 16
41
- merge_method: slerp
42
- base_model: appvoid/massive
43
- parameters:
44
- t:
45
- - filter: self_attn
46
- value:
47
- - 0
48
- - 0.5
49
- - 0.3
50
- - 0.7
51
- - 1
52
- - filter: mlp
53
- value:
54
- - 1
55
- - 0.5
56
- - 0.7
57
- - 0.3
58
- - 0
59
- - value: 0.5
60
- dtype: float16
61
- ```
 
1
+
2
+
3
  ---
4
+ language:
5
+ - en
6
+ license: apache-2.0
 
7
  tags:
8
+ - text-generation-inference
9
+ - transformers
10
+ - unsloth
11
+ - llama
12
+ - trl
13
+ - sft
14
  ---
15
+
16
+ <style>
17
+ @import url('https://fonts.googleapis.com/css2?family=Vollkorn:ital,wght@0,400..900;1,400..900&display=swap');
18
+ </style>
19
+
20
+ <div style="background-color: #101010; border-radius: .5rem; padding: 2rem; font-family: monospace; font-size: .85rem; text-align: justify;">
21
+
22
+ ![palmer-004](https://huggingface.co/appvoid/palmer-004-original/resolve/main/palmer-004.jpeg)
23
+
24
+
25
+ #### palmer turbo
26
+
27
+ This model has a slightly different architecture and training style:
28
+
29
+ 1. The model was followed by a continual pretraining (lm_head + embedding layers were tuned).
30
+ 2. Base model was trained on 15k instruction/response pairs.
31
+ 3. Similar architecture than palmer series but smaller in context size (8192)
32
+
33
+ In short, palmer is now half the size, twice the speed and same overall performance with a dramatical boost on arc challenge instead of winogrande.
34
+
35
+ As all palmer models, the model is biased to respond to answers without using any specific prompt, feel free to further fine-tune it for your specific use case.
36
+
37
+ | Model | MMLU | ARC-C | HellaSwag | PIQA | Winogrande | Average |
38
+ |--------------------------------|-------|-------|-----------|--------|------------|---------|
39
+ | tinyllama | 0.2577| 0.3029| 0.5935 | 0.7329 | 0.5959 | 0.4966 |
40
+ | palmer-004-turbo |**0.2736**|**0.3558**| 0.6031 |0.7367|**0.6117**|0.5191|
41
+ | palmer-004 | 0.2661 |0.3490 | 0.6173 | **0.7481** | **0.6417** | 0.5244|
42
+
43
+
44
+ #### thanks to
45
+
46
+ - h2oai: performant base model provider
47
+ - teknium: openhermes dataset provider
48
+ - unsloth: tooling for training software
49
+
50
+ #### note
51
+
52
+ Next versions of this model will be available through my upcoming app. Keep tuned to not miss the release date on my X account.
53
+ </div>