Triangle104 commited on
Commit
d8e2277
·
verified ·
1 Parent(s): a24878b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +82 -16
README.md CHANGED
@@ -1,24 +1,90 @@
1
- ---
2
- license: other
3
- license_name: mrl
4
- language:
5
- - en
6
- - zh
7
- base_model: allura-org/Bigger-Body-8b
8
- library_name: transformers
9
- tags:
10
- - axolotl
11
- - roleplay
12
- - conversational
13
- - chat
14
- - llama-cpp
15
- - gguf-my-repo
16
- ---
17
 
18
  # Triangle104/Bigger-Body-8b-Q8_0-GGUF
19
  This model was converted to GGUF format from [`allura-org/Bigger-Body-8b`](https://huggingface.co/allura-org/Bigger-Body-8b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
20
  Refer to the [original model card](https://huggingface.co/allura-org/Bigger-Body-8b) for more details on the model.
21
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
  ## Use with llama.cpp
23
  Install llama.cpp through brew (works on Mac and Linux)
24
 
 
1
+ ---
2
+ license: other
3
+ license_name: mrl
4
+ language:
5
+ - en
6
+ - zh
7
+ base_model: allura-org/Bigger-Body-8b
8
+ library_name: transformers
9
+ tags:
10
+ - axolotl
11
+ - roleplay
12
+ - conversational
13
+ - chat
14
+ - llama-cpp
15
+ - gguf-my-repo
16
+ ---
17
 
18
  # Triangle104/Bigger-Body-8b-Q8_0-GGUF
19
  This model was converted to GGUF format from [`allura-org/Bigger-Body-8b`](https://huggingface.co/allura-org/Bigger-Body-8b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
20
  Refer to the [original model card](https://huggingface.co/allura-org/Bigger-Body-8b) for more details on the model.
21
 
22
+ ---
23
+ A roleplay-focused pseudo full-finetune of Ministral Instruct 2410. The successor to the Ink series.
24
+
25
+
26
+
27
+
28
+
29
+
30
+
31
+ Dataset
32
+ -
33
+ The Bigger Body (referred to as Ink v2.1, because that's still the
34
+ internal name) mix is absolutely disgusting. It's even more cursed than
35
+ the original Ink mix.
36
+
37
+
38
+ (Public) Original Datasets
39
+ -
40
+ -Fizzarolli/limarp-processed
41
+
42
+ -Norquinal/OpenCAI - two_users split
43
+
44
+ -allura-org/Celeste1.x-data-mixture
45
+
46
+ -mapsila/PIPPA-ShareGPT-formatted-named
47
+
48
+ allenai/tulu-3-sft-personas-instruction-following
49
+
50
+ -readmehay/medical-01-reasoning-SFT-json
51
+
52
+ -LooksJuicy/ruozhiba
53
+
54
+ -shibing624/roleplay-zh-sharegpt-gpt4-data
55
+
56
+ -CausalLM/Retrieval-SFT-Chat
57
+
58
+ -ToastyPigeon/fujin-filtered-instruct
59
+
60
+ Recommended Settings
61
+ -
62
+
63
+ Chat template: Mistral v7-tekken (NOT v3-tekken !!!! the main difference is that v7 has specific [SYSTEM_PROMPT] and [/SYSTEM_PROMPT] tags)
64
+ Recommended samplers (not the be-all-end-all, try some on your own!):
65
+
66
+ I have literally no idea. you're on your own.
67
+
68
+ Hyperparams
69
+ -
70
+ General
71
+
72
+ Epochs = 2
73
+ LR = 2e-6
74
+ LR Scheduler = Cosine
75
+ Optimizer = Apollo-mini
76
+ Optimizer target modules = all_linear
77
+ Effective batch size = 16
78
+ Weight Decay = 0.01
79
+ Warmup steps = 50
80
+ Total steps = 920
81
+
82
+ Credits
83
+ -
84
+ Humongous thanks to the people who created the data.
85
+ Big thanks to all Allura members for testing and emotional support ilya /platonic
86
+
87
+ ---
88
  ## Use with llama.cpp
89
  Install llama.cpp through brew (works on Mac and Linux)
90