Gryphe commited on
Commit
37cebac
1 Parent(s): 2375618

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -3,19 +3,19 @@ license: other
3
  language:
4
  - en
5
  ---
6
- The Llama 2 sequel to my original experiment with gradient merges using [the following script](https://github.com/Gryphe/BlockMerge_Gradient). Its three models ([Hermes](https://huggingface.co/NousResearch/Nous-Hermes-Llama2-13b), [Chronos](https://huggingface.co/elinas/chronos-13b-v2) and [Airoboros](https://huggingface.co/jondurbin/airoboros-l2-13b-gpt4-2.0)) are almost evenly divided over the layer structure this time. Airoboros was the "wildcard model" due to its superior ability to understand complex instructions.
7
 
8
  ## Model details
9
 
10
- As always, the main objective was to create an all-round model with improved roleplaying capabilities. MythoLogic-L2 differs from its predecessor in that it focuses primarily on the understanding of instructions and personalities of complex character cards.
11
 
12
- Illustrated below are the gradients used for this specific L2 recipe;
13
 
14
- [](MythoLogic-L2.png)
15
 
16
  ## Prompt Format
17
 
18
- This model primarily uses (and wast tested with) Alpaca formatting, so for optimal model performance, use:
19
  ```
20
  ### Instruction:
21
  Your instruction or question here.
 
3
  language:
4
  - en
5
  ---
6
+ The Llama 2 sequel to my [original experiment](https://huggingface.co/Gryphe/MythoLogic-13b) with gradient merges using [the following script](https://github.com/Gryphe/BlockMerge_Gradient). Its three models ([Hermes](https://huggingface.co/NousResearch/Nous-Hermes-Llama2-13b), [Chronos](https://huggingface.co/elinas/chronos-13b-v2) and [Airoboros](https://huggingface.co/jondurbin/airoboros-l2-13b-gpt4-2.0)) are almost evenly divided over the layer structure this time. Airoboros was the "wildcard model" due to its superior ability to understand complex instructions.
7
 
8
  ## Model details
9
 
10
+ As before, the main objective was to create an all-round model with improved roleplaying capabilities. MythoLogic-L2 differs from its predecessor in that it focuses primarily on the understanding of instructions and personalities of complex character cards.
11
 
12
+ Illustrated below are the gradients used for this specific L2 recipe, with the top of the image representing layer 0 and the bottom layer 40.
13
 
14
+ ![](MythoLogic-L2.png)
15
 
16
  ## Prompt Format
17
 
18
+ This model primarily uses (and was heavily tested with) Alpaca formatting, so for optimal model performance, use:
19
  ```
20
  ### Instruction:
21
  Your instruction or question here.