Triangle104 commited on
Commit
8208f31
1 Parent(s): 81dbaa8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +38 -1
README.md CHANGED
@@ -6,12 +6,49 @@ tags:
6
  - merge
7
  - llama-cpp
8
  - gguf-my-repo
 
9
  ---
10
 
11
  # Triangle104/MN-12B-Starcannon-v3-Q4_K_S-GGUF
12
  This model was converted to GGUF format from [`nothingiisreal/MN-12B-Starcannon-v3`](https://huggingface.co/nothingiisreal/MN-12B-Starcannon-v3) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
13
  Refer to the [original model card](https://huggingface.co/nothingiisreal/MN-12B-Starcannon-v3) for more details on the model.
14
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  ## Use with llama.cpp
16
  Install llama.cpp through brew (works on Mac and Linux)
17
 
@@ -50,4 +87,4 @@ Step 3: Run inference through the main binary.
50
  or
51
  ```
52
  ./llama-server --hf-repo Triangle104/MN-12B-Starcannon-v3-Q4_K_S-GGUF --hf-file mn-12b-starcannon-v3-q4_k_s.gguf -c 2048
53
- ```
 
6
  - merge
7
  - llama-cpp
8
  - gguf-my-repo
9
+ license: apache-2.0
10
  ---
11
 
12
  # Triangle104/MN-12B-Starcannon-v3-Q4_K_S-GGUF
13
  This model was converted to GGUF format from [`nothingiisreal/MN-12B-Starcannon-v3`](https://huggingface.co/nothingiisreal/MN-12B-Starcannon-v3) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
14
  Refer to the [original model card](https://huggingface.co/nothingiisreal/MN-12B-Starcannon-v3) for more details on the model.
15
 
16
+ ---
17
+ Model details:
18
+ -
19
+ This is a merge of pre-trained language models created using mergekit.
20
+
21
+ This model was merged using the TIES merge method using nothingiisreal/MN-12B-Celeste-V1.9 as a base.
22
+
23
+ Merge Fodder
24
+ -
25
+ The following models were included in the merge:
26
+
27
+ anthracite-org/magnum-12b-v2
28
+ nothingiisreal/MN-12B-Celeste-V1.9
29
+
30
+ Configuration
31
+ -
32
+ The following YAML configuration was used to produce this model:
33
+
34
+ models:
35
+ - model: anthracite-org/magnum-12b-v2
36
+ parameters:
37
+ density: 0.3
38
+ weight: 0.5
39
+ - model: nothingiisreal/MN-12B-Celeste-V1.9
40
+ parameters:
41
+ density: 0.7
42
+ weight: 0.5
43
+
44
+ merge_method: ties
45
+ base_model: nothingiisreal/MN-12B-Celeste-V1.9
46
+ parameters:
47
+ normalize: true
48
+ int8_mask: true
49
+ dtype: bfloat16
50
+
51
+ ---
52
  ## Use with llama.cpp
53
  Install llama.cpp through brew (works on Mac and Linux)
54
 
 
87
  or
88
  ```
89
  ./llama-server --hf-repo Triangle104/MN-12B-Starcannon-v3-Q4_K_S-GGUF --hf-file mn-12b-starcannon-v3-q4_k_s.gguf -c 2048
90
+ ```