giraffe176
commited on
Commit
•
f4f39a2
1
Parent(s):
c7836d2
Update README.md
Browse files
README.md
CHANGED
@@ -23,7 +23,7 @@ As part of this process, I tried to figure out if there was a way to determine a
|
|
23 |
Way too late in the process, did I learn that [dare_ties](https://arxiv.org/abs/2311.03099) has a random element to it, but considered it valuable information for next time. After concluding that project, I began collecting more data, this time setting a specified seed in mergekit for reproducibility.
|
24 |
This model is *not* a result of the above work but is the genesis of how this model came to be.
|
25 |
|
26 |
-
I present, Starling_Monarch_Westlake_Garten-7B-v0.1
|
27 |
|
28 |
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
29 |
|
|
|
23 |
Way too late in the process, did I learn that [dare_ties](https://arxiv.org/abs/2311.03099) has a random element to it, but considered it valuable information for next time. After concluding that project, I began collecting more data, this time setting a specified seed in mergekit for reproducibility.
|
24 |
This model is *not* a result of the above work but is the genesis of how this model came to be.
|
25 |
|
26 |
+
I present, **Starling_Monarch_Westlake_Garten-7B-v0.1**, the only 7B model to score > 80 on the EQ-Bench v2.1 benchmark found [here](https://github.com/EQ-bench/EQ-Bench), outscoring larger models like [abacusai/Smaug-72B-v0.1](https://huggingface.co/abacusai/Smaug-72B-v0.1) and [cognitivecomputations/dolphin-2.2-70b](https://huggingface.co/cognitivecomputations/dolphin-2.2-70b)
|
27 |
|
28 |
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
29 |
|