grimjim commited on
Commit
3efe075
1 Parent(s): e01fdde

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +69 -67
README.md CHANGED
@@ -1,67 +1,69 @@
1
- ---
2
- base_model:
3
- - Hastagaras/Halu-OAS-8B-Llama3
4
- - openlynn/Llama-3-Soliloquy-8B-v2
5
- - grimjim/llama-3-aaditya-OpenBioLLM-8B
6
- - NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
7
- - mlabonne/NeuralDaredevil-8B-abliterated
8
- library_name: transformers
9
- tags:
10
- - mergekit
11
- - merge
12
- license: llama3
13
- license_link: LICENSE
14
- ---
15
- # Llama-3-Steerpike-v1-OAS-8B
16
-
17
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
18
-
19
- This model might result in characters who are "too" smart if conversation veers into the analytical, but that may be fine depending on the context.
20
-
21
- Tested lightly with Instruct prompts, minP=0.01, and temperature 1+.
22
-
23
- Built with Meta Llama 3.
24
-
25
- ## Merge Details
26
- ### Merge Method
27
-
28
- This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [mlabonne/NeuralDaredevil-8B-abliterated](https://huggingface.co/mlabonne/NeuralDaredevil-8B-abliterated) as a base.
29
-
30
- ### Models Merged
31
-
32
- The following models were included in the merge:
33
- * [Hastagaras/Halu-OAS-8B-Llama3](https://huggingface.co/Hastagaras/Halu-OAS-8B-Llama3)
34
- * [openlynn/Llama-3-Soliloquy-8B-v2](https://huggingface.co/openlynn/Llama-3-Soliloquy-8B-v2)
35
- * [grimjim/llama-3-aaditya-OpenBioLLM-8B](https://huggingface.co/grimjim/llama-3-aaditya-OpenBioLLM-8B)
36
- * [NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS](https://huggingface.co/NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS)
37
-
38
- ### Configuration
39
-
40
- The following YAML configuration was used to produce this model:
41
-
42
- ```yaml
43
- base_model: mlabonne/NeuralDaredevil-8B-abliterated
44
- dtype: bfloat16
45
- merge_method: task_arithmetic
46
- slices:
47
- - sources:
48
- - layer_range: [0, 32]
49
- model: mlabonne/NeuralDaredevil-8B-abliterated
50
- - layer_range: [0, 32]
51
- model: NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
52
- parameters:
53
- weight: 0.5
54
- - layer_range: [0, 32]
55
- model: Hastagaras/Halu-OAS-8B-Llama3
56
- parameters:
57
- weight: 0.2
58
- - layer_range: [0, 32]
59
- model: openlynn/Llama-3-Soliloquy-8B-v2
60
- parameters:
61
- weight: 0.03
62
- - layer_range: [0, 32]
63
- model: grimjim/llama-3-aaditya-OpenBioLLM-8B
64
- parameters:
65
- weight: 0.1
66
-
67
- ```
 
 
 
1
+ ---
2
+ base_model:
3
+ - Hastagaras/Halu-OAS-8B-Llama3
4
+ - openlynn/Llama-3-Soliloquy-8B-v2
5
+ - grimjim/llama-3-aaditya-OpenBioLLM-8B
6
+ - NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
7
+ - mlabonne/NeuralDaredevil-8B-abliterated
8
+ library_name: transformers
9
+ tags:
10
+ - mergekit
11
+ - merge
12
+ license: llama3
13
+ license_link: LICENSE
14
+ ---
15
+ # Llama-3-Steerpike-v1-OAS-8B
16
+
17
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
18
+
19
+ This model might result in characters who are "too" smart if conversation veers into the analytical, but that may be fine depending on the context.
20
+ There are issues early on with the consistency of formatting, though that will stabilize with more context.
21
+ This model is imperfect, but interesting.
22
+
23
+ Tested lightly with Instruct prompts, minP=0.01, and temperature 1+.
24
+
25
+ Built with Meta Llama 3.
26
+
27
+ ## Merge Details
28
+ ### Merge Method
29
+
30
+ This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [mlabonne/NeuralDaredevil-8B-abliterated](https://huggingface.co/mlabonne/NeuralDaredevil-8B-abliterated) as a base.
31
+
32
+ ### Models Merged
33
+
34
+ The following models were included in the merge:
35
+ * [Hastagaras/Halu-OAS-8B-Llama3](https://huggingface.co/Hastagaras/Halu-OAS-8B-Llama3)
36
+ * [openlynn/Llama-3-Soliloquy-8B-v2](https://huggingface.co/openlynn/Llama-3-Soliloquy-8B-v2)
37
+ * [grimjim/llama-3-aaditya-OpenBioLLM-8B](https://huggingface.co/grimjim/llama-3-aaditya-OpenBioLLM-8B)
38
+ * [NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS](https://huggingface.co/NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS)
39
+
40
+ ### Configuration
41
+
42
+ The following YAML configuration was used to produce this model:
43
+
44
+ ```yaml
45
+ base_model: mlabonne/NeuralDaredevil-8B-abliterated
46
+ dtype: bfloat16
47
+ merge_method: task_arithmetic
48
+ slices:
49
+ - sources:
50
+ - layer_range: [0, 32]
51
+ model: mlabonne/NeuralDaredevil-8B-abliterated
52
+ - layer_range: [0, 32]
53
+ model: NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
54
+ parameters:
55
+ weight: 0.5
56
+ - layer_range: [0, 32]
57
+ model: Hastagaras/Halu-OAS-8B-Llama3
58
+ parameters:
59
+ weight: 0.2
60
+ - layer_range: [0, 32]
61
+ model: openlynn/Llama-3-Soliloquy-8B-v2
62
+ parameters:
63
+ weight: 0.03
64
+ - layer_range: [0, 32]
65
+ model: grimjim/llama-3-aaditya-OpenBioLLM-8B
66
+ parameters:
67
+ weight: 0.1
68
+
69
+ ```