Triangle104 commited on
Commit
fad576d
1 Parent(s): 94f2d16

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +77 -1
README.md CHANGED
@@ -13,12 +13,88 @@ datasets:
13
  - anthracite-org/kalo-opus-instruct-3k-filtered-no-system
14
  - anthracite-org/nopm_claude_writing_fixed
15
  base_model: crestf411/MN-Slush
 
16
  ---
17
 
18
  # Triangle104/MN-Slush-Q8_0-GGUF
19
  This model was converted to GGUF format from [`crestf411/MN-Slush`](https://huggingface.co/crestf411/MN-Slush) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
20
  Refer to the [original model card](https://huggingface.co/crestf411/MN-Slush) for more details on the model.
21
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
  ## Use with llama.cpp
23
  Install llama.cpp through brew (works on Mac and Linux)
24
 
@@ -57,4 +133,4 @@ Step 3: Run inference through the main binary.
57
  or
58
  ```
59
  ./llama-server --hf-repo Triangle104/MN-Slush-Q8_0-GGUF --hf-file mn-slush-q8_0.gguf -c 2048
60
- ```
 
13
  - anthracite-org/kalo-opus-instruct-3k-filtered-no-system
14
  - anthracite-org/nopm_claude_writing_fixed
15
  base_model: crestf411/MN-Slush
16
+ license: apache-2.0
17
  ---
18
 
19
  # Triangle104/MN-Slush-Q8_0-GGUF
20
  This model was converted to GGUF format from [`crestf411/MN-Slush`](https://huggingface.co/crestf411/MN-Slush) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
21
  Refer to the [original model card](https://huggingface.co/crestf411/MN-Slush) for more details on the model.
22
 
23
+ ---
24
+ Model details:
25
+ -
26
+ Slush is a two-stage model trained with high LoRA
27
+ dropout, where stage 1 is a pretraining continuation on the base model,
28
+ aimed at boosting the model's creativity and writing capabilities. This
29
+ is then merged into the instruction tune model, and stage 2 is a fine
30
+ tuning step on top of this to further enhance its roleplaying
31
+ capabilities and/or to repair any damage caused in the stage 1 merge.
32
+
33
+ This is still early stage. As always, feedback is welcome, and begone if you demand perfection.
34
+
35
+ The second stage, like the Sunfall series, follows the Silly
36
+ Tavern preset (Mistral V2 & V3, though V3-Tekken works fine), so
37
+ ymmv in particular if you use some other tool and/or preset.
38
+
39
+ Parameter suggestions:
40
+ -
41
+ I did all my testing with temp 1, min-p 0.1, DRY 0.8.
42
+
43
+ Training details:
44
+ -
45
+ Stage 1 (continued pretraining)
46
+ Target: mistralai/Mistral-Nemo-Base-2407 (resulting LoRA merged into mistralai/Mistral-Nemo-Instruct-2407)
47
+ LoRA dropout 0.5 (motivation)
48
+ LoRA rank 64, alpha 128 (motivation)
49
+ LR cosine 4e-6
50
+ LoRA+ with LR Ratio: 15
51
+ Context size: 16384
52
+ Gradient accumulation steps: 4
53
+ Epochs: 1
54
+
55
+ Stage 2 (fine tune)
56
+ Target: Stage 1 model
57
+ LoRA dropout 0.5
58
+ LoRA rank 32, alpha 64
59
+ LR cosine 5e-6 (min 5e-7)
60
+ LoRA+ with LR Ratio: 15
61
+ Context size: 16384
62
+ Gradient accumulation steps: 4
63
+ Epochs: 2
64
+
65
+ Merge Method
66
+ -
67
+
68
+ This model was merged using the TIES merge method using mistralai/Mistral-Nemo-Base-2407 as a base.
69
+
70
+ Configuration
71
+ -
72
+ The following YAML configuration was used to produce this model:
73
+
74
+ models:
75
+ - model: stage1-on-instruct
76
+ parameters:
77
+ weight: 1
78
+ density: 1
79
+ - model: stage2-on-stage1
80
+ parameters:
81
+ weight: 0.7
82
+ density: 1
83
+ - model: mistralai/Mistral-Nemo-Instruct-2407
84
+ parameters:
85
+ weight: 1
86
+ density: 1
87
+ merge_method: ties
88
+ base_model: mistralai/Mistral-Nemo-Base-2407
89
+ parameters:
90
+ weight: 1
91
+ density: 1
92
+ normalize: true
93
+ int8_mask: true
94
+ tokenizer_source: mistralai/Mistral-Nemo-Instruct-2407
95
+ dtype: bfloat16
96
+
97
+ ---
98
  ## Use with llama.cpp
99
  Install llama.cpp through brew (works on Mac and Linux)
100
 
 
133
  or
134
  ```
135
  ./llama-server --hf-repo Triangle104/MN-Slush-Q8_0-GGUF --hf-file mn-slush-q8_0.gguf -c 2048
136
+ ```