DavidAhn commited on
Commit
7823ba1
1 Parent(s): dcea1df

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -10,9 +10,9 @@ base_model:
10
  - mlabonne/Daredevil-8B-abliterated
11
  ---
12
 
13
- # Llama-3-8B-slerp-262k-ties
14
 
15
- Llama-3-8B-slerp-262k-ties is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
16
  * [pankajmathur/orca_mini_v5_8b_dpo](https://huggingface.co/pankajmathur/orca_mini_v5_8b_dpo)
17
  * [mlabonne/Daredevil-8B-abliterated](https://huggingface.co/mlabonne/Daredevil-8B-abliterated)
18
 
@@ -20,7 +20,7 @@ Llama-3-8B-slerp-262k-ties is a merge of the following models using [LazyMergeki
20
 
21
  ```yaml
22
  models:
23
- - model: DavidAhn/Llama-3-8B-slerp-262k-SauerkrautLM
24
  # no parameters necessary for base model
25
  - model: pankajmathur/orca_mini_v5_8b_dpo
26
  parameters:
@@ -31,7 +31,7 @@ models:
31
  density: 0.5
32
  weight: 0.3
33
  merge_method: ties
34
- base_model: DavidAhn/Llama-3-8B-slerp-262k-SauerkrautLM
35
  parameters:
36
  normalize: true
37
  dtype: float16
@@ -46,7 +46,7 @@ from transformers import AutoTokenizer
46
  import transformers
47
  import torch
48
 
49
- model = "DavidAhn/Llama-3-8B-slerp-262k-ties"
50
  messages = [{"role": "user", "content": "What is a large language model?"}]
51
 
52
  tokenizer = AutoTokenizer.from_pretrained(model)
 
10
  - mlabonne/Daredevil-8B-abliterated
11
  ---
12
 
13
+ # Llama-3-8B-combo-merge
14
 
15
+ Llama-3-8B-combo-merge is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
16
  * [pankajmathur/orca_mini_v5_8b_dpo](https://huggingface.co/pankajmathur/orca_mini_v5_8b_dpo)
17
  * [mlabonne/Daredevil-8B-abliterated](https://huggingface.co/mlabonne/Daredevil-8B-abliterated)
18
 
 
20
 
21
  ```yaml
22
  models:
23
+ - model: DavidAhn/Llama-3-8B-combo-merge
24
  # no parameters necessary for base model
25
  - model: pankajmathur/orca_mini_v5_8b_dpo
26
  parameters:
 
31
  density: 0.5
32
  weight: 0.3
33
  merge_method: ties
34
+ base_model: DavidAhn/Llama-3-8B-combo-merge
35
  parameters:
36
  normalize: true
37
  dtype: float16
 
46
  import transformers
47
  import torch
48
 
49
+ model = "DavidAhn/Llama-3-8B-combo-merge"
50
  messages = [{"role": "user", "content": "What is a large language model?"}]
51
 
52
  tokenizer = AutoTokenizer.from_pretrained(model)