Edit model card

๐ŸŒฒ Hinoki-Sak-Sta-slerp-7B

Hinoki-Sak-Sta-slerp-7B is a merge of the following models using the LazyMergekit of Maxime Labonne powered by MergeKit of Arcee AI:

๐Ÿ’ป Configuration

slices:
  - sources:
      - model: SakanaAI/EvoLLM-JP-A-v1-7B
        layer_range: [0, 32]
      - model: stabilityai/japanese-stablelm-base-gamma-7b
        layer_range: [0, 32]
merge_method: slerp
base_model: SakanaAI/EvoLLM-JP-A-v1-7B
parameters:
  t:
    - filter: self_attn
      value: [0, 0.5, 0.3, 0.7, 1]
    - filter: mlp
      value: [1, 0.5, 0.7, 0.3, 0]
    - value: 0.5
dtype: bfloat16

๐Ÿค— Usage for HuggingFace

from transformers import AutoTokenizer, AutoModelForCausalLM
from transformers import pipeline
import torch

model_name = "AkimfromParis/Hinoki-Sak-Sta-slerp-7B"

tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.bfloat16)

pipe = pipeline("text-generation", model=model, tokenizer=tokenizer, pad_token_id=tokenizer.eos_token_id)

messages = [
    {"role": "system","content": "ใ‚ใชใŸใฏ่ช ๅฎŸใงๅ„ช็ง€ใชๆ—ฅๆœฌไบบใฎใ‚ขใ‚ทใ‚นใ‚ฟใƒณใƒˆใงใ™ใ€‚ไปฅไธ‹ใฎใƒˆใƒ”ใƒƒใ‚ฏใซ้–ขใ™ใ‚‹่ฉณ็ดฐใชๆƒ…ๅ ฑใ‚’ๆไพ›ใ—ใฆใใ ใ•ใ„ใ€‚"},
    {"role": "user", "content": "ๅคง่ฐท็ฟ”ๅนณ้ธๆ‰‹ใฏ่ชฐใงใ™ใ‹๏ผŸ"},
    ]
print(pipe(messages, max_new_tokens=512)[0]['generated_text'][-1])

๐Ÿ”– Citation

@misc{goddard2024arcee,
  title={Arcee's MergeKit: A Toolkit for Merging Large Language Models},
  author={Goddard, Charles and Siriwardhana, Shamane and Ehghaghi, Malikeh and Meyers, Luke and Karpukhin, Vlad and Benedict, Brian and McQuade, Mark and Solawetz, Jacob},
  journal={arXiv preprint arXiv:2403.13257},
  year={2024}
}

arxiv.org/abs/2403.13257

Downloads last month
0
Safetensors
Model size
7.24B params
Tensor type
BF16
ยท
Inference API
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.

Merge of