ChuckMcSneed commited on
Commit
26b3ed1
1 Parent(s): b54fd6d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -128,7 +128,7 @@ Thanks for GGUF, [@Nexesenex](https://huggingface.co/Nexesenex)!
128
  | P | 5.5 | 2.25 |
129
  | Total | 14.75 | 8.25 |
130
 
131
- ### Open LLM leaderboard
132
  [Leaderboard on Huggingface](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
133
  |Model |Average|ARC |HellaSwag|MMLU |TruthfulQA|Winogrande|GSM8K|
134
  |---------------------------------------|-------|-----|---------|-----|----------|----------|-----|
@@ -137,7 +137,7 @@ Thanks for GGUF, [@Nexesenex](https://huggingface.co/Nexesenex)!
137
  |Difference |3.83 |1.62 |0.99 |4.7 |1.89 |2.44 |11.3 |
138
 
139
  Here the losses seem far less brutal than on my bench. It seems that extending with LongLORA kills MMLU and GSM8K performance.
140
- # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
141
  Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_ChuckMcSneed__WinterGoddess-1.4x-70b-32k)
142
 
143
  | Metric |Value|
 
128
  | P | 5.5 | 2.25 |
129
  | Total | 14.75 | 8.25 |
130
 
131
+ ### [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
132
  [Leaderboard on Huggingface](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
133
  |Model |Average|ARC |HellaSwag|MMLU |TruthfulQA|Winogrande|GSM8K|
134
  |---------------------------------------|-------|-----|---------|-----|----------|----------|-----|
 
137
  |Difference |3.83 |1.62 |0.99 |4.7 |1.89 |2.44 |11.3 |
138
 
139
  Here the losses seem far less brutal than on my bench. It seems that extending with LongLORA kills MMLU and GSM8K performance.
140
+
141
  Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_ChuckMcSneed__WinterGoddess-1.4x-70b-32k)
142
 
143
  | Metric |Value|