Oumuamua-7b-instruct-GGUF
This is quantized version of nitky/Oumuamua-7b-instruct created using llama.cpp
Model Description
This is a merge of pre-trained language models created using mergekit.
Output example
[INST] <<SYS>>
ใใชใใฏๆฅๆฌ่ชใ่ฉฑใๅช็งใชใขใทในใฟใณใใงใใๅ็ญใซใฏๅฟ
ใๆฅๆฌ่ชใง็ญใใฆใใ ใใใใพใ่ใใ้็จใๅบๅใใฆใใ ใใใ
<</SYS>>
้่ทฏใงๆณฃใใฆใใๅญไพใใใพใใใใฎๅญไพใซใฏไฝใใใฃใใจ่ใใใใพใใ๏ผ [/INST] ้่ทฏใงๆณฃใใฆใใๅญไพใใใๅ ดๅใ่ใใใใ็็ฑใฏใใใคใใใใพใใ
1. ๆชๆใใใ: ่ปขใใงๆชๆใใใใใไฝใใซใถใคใใฃใฆ็ใฟใๆใใฆใใๅฏ่ฝๆงใใใใพใใ
2. ่ฟทๅญใซใชใฃใ: ่ฆชใไฟ่ญท่
ใจใฏใใใฆใใพใฃใใใพใใฏ้ใซ่ฟทใฃใฆไธๅฎใซใชใฃใฆใใๅฏ่ฝๆงใใใใพใใ
3. ใ่
นใ็ฉบใใ: ้ทๆ้้ฃไบใใใฆใใชใใใพใใฏใ่
นใ็ฉบใใฆใใใใใซๆณฃใใฆใใๅฏ่ฝๆงใใใใพใใ
4. ็ฒใใ: ้ทๆ้ๆญฉใใฆใใใใ้ใณ็ฒใใฆ็ฒใใฆใใๅฏ่ฝๆงใใใใพใใ
5. ๅฏใใใฃใ: ไธไบบใงใใใใจใซๅฏใใใๆใใฆใใใใพใใฏ่ฆชใๅ้ใใใชใใฆๅฏใใใจๆใใฆใใๅฏ่ฝๆงใใใใพใใ
6. ๆใๆใใใใ: ่ปใไธๅฏฉ่
ใซ้ญ้ใใใใพใใฏไฝใใซ้ฉใใๅฏ่ฝๆงใใใใพใใ
7. ไฝ่ชฟใๆชใ: ็บ็ฑใ้ ญ็ใๅใๆฐใชใฉใฎไฝ่ชฟไธ่ฏใๆใใฆใใๅฏ่ฝๆงใใใใพใใ
8. ใใคใฌใซ่กใใใ: ใใใคใๆฟกใใฆใใใใใใคใฌใซ่กใใใใฆๆณฃใใฆใใๅฏ่ฝๆงใใใใพใใ
9. ไฝใใๆฑใใฆใใ: ๆณจๆใๅผใใใใฃใใใๅฉใใๆฑใใฆใใๅฏ่ฝๆงใใใใพใใ
ๅญไพใฎ็ถๆณใใใ่ฆณๅฏใใๅฃฐใใใใฆ่ฉฑใ่ใใฆใฟใใใจใๅคงๅใงใใๅฟ
่ฆใซๅฟใใฆใไฟ่ญท่
ใ่ญฆๅฏใๆๆฅ่ปใๅผใถใชใฉใฎๅฏพๅฟใใใๅฟ
่ฆใใใใใใใใพใใใ
Test environment
This model was tested using text-generation-webui. I use preset min_p
and Null preset
with temperature=0.3 for Generation.
Usage
This format must be adhered to strictly, as deviations may result in less optimal outputs from the model.
The template used to construct a prompt for the Instruct model is specified as follows:
<s>[INST] <<SYS>>\n{SYSTEM_PROMPT}\n<</SYS>>\n\n{USER_MESSAGE_1} [/INST] {BOT_MESSAGE_1}</s>[INST] {USER_MESSAGE_2} [/INST]
Please be aware that <s>
and </s>
are special tokens used for the beginning of string (BOS) and end of string (EOS), respectively, while [INST] and [/INST] are considered regular strings.
For the "{SYSTEM_PROMPT}" part, We recommend using "ใใชใใฏๆฅๆฌ่ชใ่ฉฑใๅช็งใชใขใทในใฟใณใใงใใๅ็ญใซใฏๅฟ ใๆฅๆฌ่ชใง็ญใใฆใใ ใใใใพใ่ใใ้็จใๅบๅใใฆใใ ใใใ"
For the "{USER_MESSAGE_1}" part, We recommend using {instruction}\n{input}
In other words, We recommend the following:
<s>[INST] <<SYS>>\nใใชใใฏๆฅๆฌ่ชใ่ฉฑใๅช็งใชใขใทในใฟใณใใงใใๅ็ญใซใฏๅฟ
ใๆฅๆฌ่ชใง็ญใใฆใใ ใใใใพใ่ใใ้็จใๅบๅใใฆใใ ใใใ\n<</SYS>>\n\n{instruction1}\n{input1} [/INST] {BOT_MESSAGE_1}</s>[INST] {instruction2}\n{input2} [/INST]
Use the instruct model
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
model_name = "nitky/Oumuamua-7b-instruct"
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.bfloat16, device_map="auto")
tokenizer = AutoTokenizer.from_pretrained(model_name)
device = "cuda"
messages = [
{"role": "system", "content": "ใใชใใฏๆฅๆฌ่ชใ่ฉฑใๅช็งใชใขใทในใฟใณใใงใใๅ็ญใซใฏๅฟ
ใๆฅๆฌ่ชใง็ญใใฆใใ ใใใใพใ่ใใ้็จใๅบๅใใฆใใ ใใใ"},
{"role": "user", "content": "้่ทฏใงๆณฃใใฆใใๅญไพใใใพใใใใฎๅญไพใซใฏไฝใใใฃใใจ่ใใใใพใใ๏ผ"}
]
encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt")
model_inputs = encodeds.to(device)
model.to(device)
generated_ids = model.generate(model_inputs, max_new_tokens=256, do_sample=True, temperature=0.3)
decoded = tokenizer.batch_decode(generated_ids)
print(decoded[0])
Merge Details
Merge Method
This model was merged using the Model Stock merge method using nitky/Oumuamua-7b-base as a base.
Models Merged
The following models were included in the merge:
- stabilityai/japanese-stablelm-base-gamma-7b
- tokyotech-llm/Swallow-MS-7b-v0.1
- DataPilot/ArrowPro-7B-KillerWhale
- nitky/RP-7b-instruct
- mistralai/Mistral-7B-Instruct-v0.3
- mistralai/Mistral-7B-v0.1
Configuration
The following YAML configuration was used to produce this model:
merge_method: task_arithmetic
base_model: mistralai/Mistral-7B-v0.1
models:
- model: tokyotech-llm/Swallow-MS-7b-v0.1
parameters:
weight:
- filter: embed_tokens
value: 1.0
- value: 0
dtype: bfloat16
tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1
name: Mistral-7B-v0.1-VE-Swallow-MS
---
merge_method: task_arithmetic
base_model: stabilityai/japanese-stablelm-base-gamma-7b
models:
- model: tokyotech-llm/Swallow-MS-7b-v0.1
parameters:
weight:
- filter: embed_tokens
value: 1.0
- value: 0
dtype: bfloat16
tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1
name: japanese-stablelm-base-gamma-7b-VE-Swallow-MS
---
merge_method: task_arithmetic
base_model: DataPilot/ArrowPro-7B-KillerWhale
models:
- model: tokyotech-llm/Swallow-MS-7b-v0.1
parameters:
weight:
- filter: embed_tokens
value: 1.0
- value: 0
dtype: bfloat16
tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1
name: ArrowPro-7B-KillerWhale-VE-Swallow-MS
---
merge_method: task_arithmetic
base_model: nitky/RP-7b-instruct
models:
- model: tokyotech-llm/Swallow-MS-7b-v0.1
parameters:
weight:
- filter: embed_tokens
value: 1.0
- value: 0
dtype: bfloat16
tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1
name: RP-7b-instruct-VE-Swallow-MS
---
merge_method: task_arithmetic
base_model: mistralai/Mistral-7B-Instruct-v0.3
models:
- model: tokyotech-llm/Swallow-MS-7b-v0.1
parameters:
weight:
- filter: embed_tokens
value: 1.0
- value: 0
dtype: bfloat16
tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1
name: Mistral-7B-Instruct-v0.3-VE-Swallow-MS
---
merge_method: linear
models:
- model: nitky/Oumuamua-7b-base
parameters:
weight: 1.0
- model: tokyotech-llm/Swallow-MS-7b-instruct-v0.1
parameters:
weight: 0.8
- model: tokyotech-llm/Swallow-MS-7b-v0.1
parameters:
weight: -0.8
- model: Mistral-7B-Instruct-v0.3-VE-Swallow-MS
parameters:
weight: 0.8
- model: Mistral-7B-v0.1-VE-Swallow-MS
parameters:
weight: -0.8
dtype: bfloat16
name: Oumuamua-7b-instruct-alpha
---
merge_method: linear
models:
- model: nitky/Oumuamua-7b-base
parameters:
weight: 1.0
- model: ArrowPro-7B-KillerWhale-VE-Swallow-MS # include Mistral-7B-Instruct-v0.3
parameters:
weight: 0.8
- model: japanese-stablelm-base-gamma-7b-VE-Swallow-MS
parameters:
weight: -0.8
dtype: bfloat16
name: Oumuamua-7b-instruct-beta
---
merge_method: linear
models:
- model: nitky/Oumuamua-7b-base
parameters:
weight: 1.0
- model: RP-7b-instruct-VE-Swallow-MS
parameters:
weight: 0.8
- model: japanese-stablelm-base-gamma-7b-VE-Swallow-MS
parameters:
weight: -0.8
- model: Mistral-7B-Instruct-v0.3-VE-Swallow-MS
parameters:
weight: 0.8
- model: Mistral-7B-v0.1-VE-Swallow-MS
parameters:
weight: -0.8
dtype: bfloat16
name: Oumuamua-7b-instruct-gamma
---
merge_method: model_stock
base_model: nitky/Oumuamua-7b-base
models:
- model: Oumuamua-7b-instruct-alpha
- model: Oumuamua-7b-instruct-beta
- model: Oumuamua-7b-instruct-gamma
dtype: bfloat16
name: Oumuamua-7b-instruct
- Downloads last month
- 83
Model tree for QuantFactory/Oumuamua-7b-instruct-GGUF
Base model
nitky/Oumuamua-7b-instruct