Joseph717171 commited on
Commit
6594260
1 Parent(s): fcbadf2

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +69 -0
README.md ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: []
3
+ library_name: transformers
4
+ tags:
5
+ - mergekit
6
+ - merge
7
+
8
+ ---
9
+ # Cerebrum-1.0-12.25B
10
+
11
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
12
+
13
+ ## Merge Details
14
+ ### Merge Method
15
+
16
+ This model was merged using the passthrough merge method.
17
+
18
+ ### Models Merged
19
+
20
+ The following models were included in the merge:
21
+ * /Users/jsarnecki/opt/Workspace/AetherResearch/Cerebrum-1.0-7b
22
+
23
+ ### Configuration
24
+
25
+ The following YAML configuration was used to produce this model:
26
+
27
+ ```yaml
28
+ dtype: bfloat16
29
+ merge_method: passthrough
30
+ # Layer-Interleaving of AetherResearch/Cerebrum-1.0-7b
31
+ # where n = 32 (The number of layers the model originally had)
32
+ # m = 8 (The number of layers to interleave per row)
33
+ # M = 24 = 3m (The number of layers interleaved into the model)
34
+ # s = 56 = 32 + 24 = n + M (The number of layers the model will have after the layer interleaving)
35
+ # P = 12.25B parameters = (s/n)*7 (The number of parameters the model will have after the layer interleaving)
36
+ slices:
37
+ - sources:
38
+ - layer_range: [0, 8]
39
+ model: /Users/jsarnecki/opt/Workspace/AetherResearch/Cerebrum-1.0-7b
40
+
41
+ # Interleaved Layer
42
+ - sources:
43
+ - layer_range: [4, 12]
44
+ model: /Users/jsarnecki/opt/Workspace/AetherResearch/Cerebrum-1.0-7b
45
+
46
+ - sources:
47
+ - layer_range: [8, 16]
48
+ model: /Users/jsarnecki/opt/Workspace/AetherResearch/Cerebrum-1.0-7b
49
+
50
+ # Interleaved Layer
51
+ - sources:
52
+ - layer_range: [12, 20]
53
+ model: /Users/jsarnecki/opt/Workspace/AetherResearch/Cerebrum-1.0-7b
54
+
55
+ - sources:
56
+ - layer_range: [16, 24]
57
+ model: /Users/jsarnecki/opt/Workspace/AetherResearch/Cerebrum-1.0-7b
58
+
59
+ # Interleaved Layer
60
+ - sources:
61
+ - layer_range: [20, 28]
62
+ model: /Users/jsarnecki/opt/Workspace/AetherResearch/Cerebrum-1.0-7b
63
+
64
+ - sources:
65
+ - layer_range: [24, 32]
66
+ model: /Users/jsarnecki/opt/Workspace/AetherResearch/Cerebrum-1.0-7b
67
+
68
+
69
+ ```