Sumail commited on
Commit
579264a
1 Parent(s): bfea208

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  base_model:
3
- - denisman/llama-2.4
4
- - denisman/llama-4.25-k-11
5
  library_name: transformers
6
  tags:
7
  - mergekit
@@ -20,8 +20,8 @@ This model was merged using the SLERP merge method.
20
  ### Models Merged
21
 
22
  The following models were included in the merge:
23
- * [denisman/llama-2.4](https://huggingface.co/denisman/llama-2.4)
24
- * [denisman/llama-4.25-k-11](https://huggingface.co/denisman/llama-4.25-k-11)
25
 
26
  ### Configuration
27
 
@@ -32,12 +32,12 @@ The following YAML configuration was used to produce this model:
32
 
33
  slices:
34
  - sources:
35
- - model: denisman/llama-4.25-k-11
36
  layer_range: [0, 48]
37
- - model: denisman/llama-2.4
38
  layer_range: [0, 48]
39
  merge_method: slerp
40
- base_model: denisman/llama-2.4
41
  parameters:
42
  t:
43
  - filter: self_attn
 
1
  ---
2
  base_model:
3
+ - 0x0grandpa0/melancholyson2
4
+ - 0x0grandpa0/melancholysdaughter
5
  library_name: transformers
6
  tags:
7
  - mergekit
 
20
  ### Models Merged
21
 
22
  The following models were included in the merge:
23
+ * [0x0grandpa0/melancholyson2](https://huggingface.co/0x0grandpa0/melancholyson2)
24
+ * [0x0grandpa0/melancholysdaughter](https://huggingface.co/0x0grandpa0/melancholysdaughter)
25
 
26
  ### Configuration
27
 
 
32
 
33
  slices:
34
  - sources:
35
+ - model: 0x0grandpa0/melancholyson2
36
  layer_range: [0, 48]
37
+ - model: 0x0grandpa0/melancholysdaughter
38
  layer_range: [0, 48]
39
  merge_method: slerp
40
+ base_model: 0x0grandpa0/melancholyson2
41
  parameters:
42
  t:
43
  - filter: self_attn
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "denisman/llama-2.4",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "0x0grandpa0/melancholyson2",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
mergekit_config.yml CHANGED
@@ -2,12 +2,12 @@
2
 
3
  slices:
4
  - sources:
5
- - model: denisman/llama-4.25-k-11
6
  layer_range: [0, 48]
7
- - model: denisman/llama-2.4
8
  layer_range: [0, 48]
9
  merge_method: slerp
10
- base_model: denisman/llama-2.4
11
  parameters:
12
  t:
13
  - filter: self_attn
 
2
 
3
  slices:
4
  - sources:
5
+ - model: 0x0grandpa0/melancholyson2
6
  layer_range: [0, 48]
7
+ - model: 0x0grandpa0/melancholysdaughter
8
  layer_range: [0, 48]
9
  merge_method: slerp
10
+ base_model: 0x0grandpa0/melancholyson2
11
  parameters:
12
  t:
13
  - filter: self_attn
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9b31ad67bde27876356d95d98caf08d049b576fb4311a23d18dc109cd9e1881f
3
  size 4945284816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:980d669bb25cbdfb355a88eb0e3986e110b2175c53e2bc623e4094c5a865766c
3
  size 4945284816
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5864f1ce460c54e7972f1f5a710ad7f9a63c3cfd384b577a328386b33089dd0c
3
  size 4934842800
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0fe545139a3aca0683a8a6368597ec2efd933bf0bd4e3b383a8b4dc20d0b765
3
  size 4934842800
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2afe7d4a9c59eec1e9af96e8f4d9ac74e1500032ba373ad31bc4e5b1e472d122
3
  size 4972600080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fb0ea5479463939ac4975da99cb454187a54685ca5c1f27537327aff3825722
3
  size 4972600080
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:60ef0d95e566d8bed6be8e0f039136488160adb5aba43d249fbba17c20a979db
3
  size 2806137296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2ce3b454e565f50c226347cafc4740529550828ba8e338627ddbfbd3ec49631
3
  size 2806137296