brucethemoose's picture
Update README.md
42550fc
|
raw
history blame
1.6 kB

-- license: other license_name: yi-license license_link: https://huggingface.co/01-ai/Yi-34B/blob/main/LICENSE language: - en library_name: transformers pipeline_tag: text-generation

NousResearch/Nous-Capybara-34B and migtissera/Tess-M-v1.2 merged with an expermental implementation of "dare" via mergekit. See:

Language Models are Super Mario: Absorbing Abilities from Homologous Models as a Free Lunch

https://arxiv.org/abs/2311.03099

https://github.com/cg123/mergekit/tree/dare-tokenizer

Merged with the following config, and the tokenizer from Yi Llamafied:

models:
  - model: /home/alpha/Storage/Models/Raw/larryvrh_Yi-34B-200K-Llamafied
    # no parameters necessary for base model
  - model: /home/alpha/Storage/Models/Raw/migtissera_Tess-M-v1.2
    parameters:
      weight: 0.62
      density: 0.55
  - model: /home/alpha/Storage/Models/Raw/Nous-Capybara-34B
    parameters:
      weight: 0.56
      density: 0.55
merge_method: dare_ties
base_model: /home/alpha/Storage/Models/Raw/larryvrh_Yi-34B-200K-Llamafied
parameters:
  int8_mask: true
dtype: bfloat16

Prompt template: Orca-Vicuna

SYSTEM: {system_message}
USER: {prompt}
ASSISTANT:

Sometimes the model "spells out" the stop token as </s> like Capybara, so you may need to add </s> as an additional stopping condition.


Credits:

https://github.com/cg123/mergekit/tree/dare-tokenizer

https://huggingface.co/NousResearch/Nous-Capybara-34B/

https://huggingface.co/migtissera/Tess-M-v1.2

https://huggingface.co/larryvrh/Yi-34B-200K-Llamafied

https://huggingface.co/01-ai/Yi-34B-200K