|
--- |
|
tags: |
|
- llama |
|
- alpaca |
|
- vicuna |
|
- uncensored |
|
- cot |
|
- chain of thought |
|
- story |
|
- adventure |
|
- roleplay |
|
- rp |
|
- merge |
|
- mix |
|
- instruct |
|
- wizardlm |
|
- superhot |
|
- supercot |
|
- manticore |
|
- hippogriff |
|
--- |
|
|
|
## 30B-Epsilon |
|
|
|
Epsilon is an instruct based general purpose model assembled from hand picked models and LoRAs. |
|
There is no censorship and it follows instructions in the Alpaca format. This means you can create |
|
your own rules in the context memory of your inference system of choice [mainly KoboldAI or Text |
|
Generation Webui and chat UIs like SillyTavern and so on]. |
|
|
|
## Composition: |
|
|
|
This model is the result of an experimental use of LoRAs on language models and model merges. |
|
[] = applied as LoRA to a composite model | () = combined as composite models |
|
30B-Epsilon = [SuperCOT[SuperHOT-prototype13b-8192[(wizardlmuncensored+((hippogriff+manticore)+(StoryV2))] |
|
|
|
Alpaca's instruct format can be used to do many things, including control of the terms of behavior |
|
between a user and a response from an agent in chat. Below is an example of a command injected into |
|
memory. |
|
|
|
``` |
|
### Instruction: |
|
Make Narrator function as a text based adventure game that responds with verbose, detailed, and creative descriptions of what happens next after Player's response. |
|
Make Player function as the player input for Narrator's text based adventure game, controlling a character named (insert character name here, their short bio, and |
|
whatever quest or other information to keep consistent in the interaction). |
|
|
|
### Response: |
|
{an empty new line here} |
|
``` |
|
|
|
All datasets from all models and LoRAs used were documented and reviewed as model candidates for merging. |
|
Model candidates were based on five core principles: creativity, logic, inference, instruction following, |
|
and longevity of trained responses. SuperHOT-prototype30b-8192 was used in this mix, not the 8K version; |
|
the prototype LoRA seems to have been removed [from HF] as of this writing. The GPT4Alpaca LoRA from |
|
Chansung was removed from this amalgam following a thorough review of where censorship and railroading |
|
the user came from in 33B-Lazarus. This is not a reflection of ChanSung's excellent work - it merely did |
|
not fit the purpose of this model. |
|
|
|
## Language Models and LoRAs Used Credits: |
|
|
|
manticore-30b-chat-pyg-alpha [Epoch0.4] by openaccess-ai-collective |
|
|
|
https://huggingface.co/openaccess-ai-collective/manticore-30b-chat-pyg-alpha |
|
|
|
hippogriff-30b-chat by openaccess-ai-collective |
|
|
|
https://huggingface.co/openaccess-ai-collective/hippogriff-30b-chat |
|
|
|
WizardLM-33B-V1.0-Uncensored by ehartford |
|
|
|
https://huggingface.co/ehartford/WizardLM-33B-V1.0-Uncensored |
|
|
|
Storytelling-LLaMa-LoRA [30B, Version 2] by GamerUnTouch |
|
|
|
https://huggingface.co/GamerUntouch/Storytelling-LLaMa-LoRAs |
|
|
|
SuperCOT-LoRA [30B] by kaiokendev |
|
|
|
https://huggingface.co/kaiokendev/SuperCOT-LoRA |
|
|
|
SuperHOT-LoRA-prototype30b-8192 [30b, not 8K version, but a removed prototype] by kaiokendev |
|
|
|
https://huggingface.co/kaiokendev/superhot-30b-8k-no-rlhf-test [Similar LoRA to one since removed that was used in making this model.] |
|
|
|
Also thanks to Meta for LLaMA and to each and every one of you |
|
who developed these fine-tunes and LoRAs. |
|
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) |
|
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_CalderaAI__30B-Epsilon) |
|
|
|
| Metric | Value | |
|
|-----------------------|---------------------------| |
|
| Avg. | 55.25 | |
|
| ARC (25-shot) | 63.05 | |
|
| HellaSwag (10-shot) | 83.59 | |
|
| MMLU (5-shot) | 56.89 | |
|
| TruthfulQA (0-shot) | 59.03 | |
|
| Winogrande (5-shot) | 77.66 | |
|
| GSM8K (5-shot) | 10.69 | |
|
| DROP (3-shot) | 35.82 | |
|
|