Edit model card

Llama-3-Kor-BCCard-Finance-8B

Llama-3-BCCard-Finance-8B is a merge of the following models using mergekit with slerp.

Dataset

๐Ÿงฉ Configuration

slices:
  - sources:
      - model: sh2orc/Llama-3-kor-BCCard-8B
        layer_range: [0, 32]
      - model: sh2orc/Llama-3-Korean-8B
        layer_range: [0, 32]
merge_method: slerp
base_model: sh2orc/Llama-3-kor-BCCard-8B
parameters:
  t:
    - filter: self_attn
      value: [0, 0.5, 0.3, 0.7, 1]
    - filter: mlp
      value: [1, 0.5, 0.7, 0.3, 0]
    - value: 0.5
  embed_slerp: true
dtype: bfloat16

๐Ÿ’ป Usage (tranformers)

!pip install -qU transformers accelerate

from transformers import AutoTokenizer
import transformers
import torch

model = "sh2orc/Llama-3-Kor-BCCard-Finance-8B"
messages = [{"role": "user", "content": "What is a large language model?"}]

tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
    "text-generation",
    model=model,
    torch_dtype=torch.float16,
    device_map="auto",
)

outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])

๐Ÿ’ป Usage (vllm)

!pip install vllm flash-attn -qU
!pip install transformers -Uq

from vllm import LLM, SamplingParams
from transformers import AutoTokenizer, pipeline

BASE_MODEL = "sh2orc/Llama-3-Kor-BCCard-Finance-8B"

llm = LLM(model=BASE_MODEL)

tokenizer = AutoTokenizer.from_pretrained(BASE_MODEL)
tokenizer.pad_token = tokenizer.eos_token
tokenizer.padding_side = 'right'

def gen(instruction):
  messages = [
      {
        "role": "system",
        "content": "๋‹น์‹ ์€ ํ›Œ๋ฅญํ•œ AI ๋น„์„œ์ž…๋‹ˆ๋‹ค. You are a great AI assistant."
      },
      {
        "role": "user",
        "content": instruction
      },
  ]

  prompt_message = tokenizer.apply_chat_template(
          messages,
          tokenize=False,
          add_generation_prompt=True,
  )

  eos_token_id = [tokenizer.eos_token_id, tokenizer.convert_tokens_to_ids("<|eot_id|>")]

  outputs = llm.generate(prompt_message, SamplingParams(stop_token_ids=eos_token_id, temperature=0.2, top_p=0.9, max_tokens=4096))

  for output in outputs:
      propt = output.prompt
      generated_text = output.outputs[0].text
      print(generated_text)

Q&A

gen('์•ˆ๋…•ํ•˜์„ธ์š”. ์ €๋Š” ์˜ค๋Š˜๋„ ํ–‰๋ณตํ•œ ํ•˜๋ฃจ๋ฅผ ๋ณด๋‚ด๊ณ  ์žˆ์Šต๋‹ˆ๋‹ค. ์ด ๋ฌธ์žฅ์„ ์˜์–ด๋กœ ๋ฒˆ์—ญํ•ด์ค˜')
Hello. I'm having a happy day today.

gen('๋‹ค์Œ ๋ฌธ์žฅ์„ ์˜์–ด๋กœ ๋ฒˆ์—ญํ•ด์ค˜. ์•ˆ๋…•ํ•˜์„ธ์š” BC์ปดํผ๋‹ˆ์ž…๋‹ˆ๋‹ค. ์˜ค๋Š˜ ์†Œ๊ฐœํ•ด๋“œ๋ฆด ์„œ๋น„์Šค๋Š” BC๊ธˆ์œต์ง€์‹ ์„œ๋น„์Šค์ž…๋‹ˆ๋‹ค.')
Hello, this is BC Company. Today, I'd like to introduce you to the BC Financial Knowledge Service.

gen('2024๋…„ ๊ธฐ์ค€ ํ‰๊ท ์†Œ๋“์ด ์–ผ๋งˆ์ด๊ณ , ์–ด๋–ค ํˆฌ์ž๋ฅผ ํ•˜๋Š” ๊ฒƒ์ด ์ข‹์€์ง€ ์•Œ๋ ค์ค˜')
2024๋…„ ๊ธฐ์ค€ ํ•œ๊ตญ ํ‰๊ท  ์›”๊ธ‰์€ ์•ฝ 4,730,000์›์ž…๋‹ˆ๋‹ค. ์ด ์ˆ˜์ต์— ๋”ฐ๋ผ ํˆฌ์ž ์ „๋žต์„ ์„ธ์šฐ๋Š” ๊ฒƒ์ด ์ข‹์Šต๋‹ˆ๋‹ค. ์˜ˆ๋ฅผ ๋“ค์–ด, 4,730,000์›์„ ํ™œ์šฉํ•˜์—ฌ ๋ถ„์‚ฐ ํˆฌ์žํ•˜๋Š” ๊ฒƒ์„ ๊ถŒ์žฅํ•ฉ๋‹ˆ๋‹ค. ์˜ˆ๋ฅผ ๋“ค์–ด, 2,000๋งŒ์›์„ ์ฃผ์‹์—, 1,000๋งŒ์›์„ ETF์—, 1,000๋งŒ์›์„ ์ฑ„๊ถŒ์—, 500๋งŒ์›์„ ์˜ˆ์ ๊ธˆ์— ํˆฌ์žํ•˜๋Š” ๋“ฑ์˜ ๋ฐฉ์‹์ž…๋‹ˆ๋‹ค. ์ด๋ ‡๊ฒŒ ํ•˜๋ฉด ๋ฆฌ์Šคํฌ๋ฅผ ๋ถ„์‚ฐํ•˜๋ฉด์„œ ์•ˆ์ •์ ์ธ ์ˆ˜์ต์„ ์ถ”๊ตฌํ•  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.

gen('2024๋…„ ๊ธฐ์ค€ ๋ฏธ๊ตญ ์ฃผ์‹ ํˆฌ์ž๋Š” ์–ด๋•Œ?')
2024๋…„ ๋ฏธ๊ตญ ์ฃผ์‹ ์‹œ์žฅ์€ ๊ธฐ์ˆ , ๊ธˆ์œต, ์—๋„ˆ์ง€ ๋“ฑ ์ผ๋ถ€ ์‚ฐ์—…์ด ๊ฐ•์„ธ๋ฅผ ๋ณด์ด๊ณ  ์žˆ์œผ๋ฉฐ, ์ด์ต ํšŒ๋ณต์ด ์˜ˆ์ƒ๋ฉ๋‹ˆ๋‹ค. ๋‹ค๋งŒ, ๊ธˆ๋ฆฌ ์ธ์ƒ๊ณผ-trade ์ „์Ÿ ๋“ฑ์˜ ๋ณ€์ˆ˜๊ฐ€ ์กด์žฌํ•˜๋ฏ€๋กœ ์‹ ์ค‘ํ•œ ํˆฌ์ž๊ฐ€ ํ•„์š”ํ•ฉ๋‹ˆ๋‹ค.

gen('2024๋…„ ๋ฏธ๊ตญ ์ฃผ์‹ ํˆฌ์ž์™€ ์ฑ„๊ถŒ ํˆฌ์ž์— ๋Œ€ํ•ด์„œ ์˜๊ฒฌ์„ ์ œ์‹œํ•ด์ค˜')
2024๋…„ ๋ฏธ๊ตญ ์ฃผ์‹ ์‹œ์žฅ์€ ์„ฑ์žฅ ์ž ์žฌ๋ ฅ์ด ๋†’์€ ์ข…๋ชฉ์— ๋Œ€ํ•œ ํˆฌ์ž๊ฐ€ ์ฃผ๋กœ ์ด๋ฃจ์–ด์งˆ ๊ฐ€๋Šฅ์„ฑ์ด ๋†’์Šต๋‹ˆ๋‹ค. ํŠนํžˆ ๊ธฐ์ˆ , ํ—ฌ์Šค์ผ€์–ด, ์žฌ์ƒ์—๋„ˆ์ง€ ๋“ฑ ๋ถ„์•ผ์˜ ์ฃผ์‹์ด ํ–ฅํ›„ ์„ฑ์žฅ ๊ฐ€๋Šฅ์„ฑ์„ ๊ณ ๋ คํ•˜์—ฌ ๋งค์ˆ˜๋  ๊ฒƒ์ž…๋‹ˆ๋‹ค. ๋˜ํ•œ, ์ธํ”Œ๋ ˆ์ด์…˜ ์™„ํ™”์™€ ํ•จ๊ป˜ ๊ฐ€์น˜์ฃผ ํˆฌ์ž๊ฐ€ ์žฌ๊ฐœ๋  ๊ฐ€๋Šฅ์„ฑ์ด ์žˆ์Šต๋‹ˆ๋‹ค.

๋ฐ˜๋ฉด, ๋ฏธ๊ตญ ์ฑ„๊ถŒ ์‹œ์žฅ์—์„œ๋Š” ์—ฐ๊ธˆ ์ˆ˜์ต๋ฅ  ์ธํ•˜์™€ ์ •๋ถ€ ๋ถ€์ฑ„ ์ฆ๊ฐ€๋กœ ์ธํ•ด ๋น„์šฉ ๋ถ€๋‹ด์ด ์ฆ๊ฐ€ํ•  ๊ฒƒ์œผ๋กœ ์˜ˆ์ƒ๋ฉ๋‹ˆ๋‹ค. ํŠนํžˆ, ์žฅ๊ธฐ ์ฑ„๊ถŒ์˜ ์ƒ๋Œ€์ ์ธ ๋น„ํ˜ธ์žฌ์ ์ธ ํ™˜๊ฒฝ์ด ์˜ˆ์ƒ๋ฉ๋‹ˆ๋‹ค.

gen('ํ•ด์™ธ์—ฌํ–‰์„ ๋งŽ์ด ๋‹ค๋‹ˆ๋Š”๋ฐ ํ˜œํƒ์ด ์ข‹์€ ์นด๋“œ์ƒํ’ˆ์„ ์ถ”์ฒœํ•ด์ค˜')
BC ๋ฐ”๋กœ ์—์–ด ํ”Œ๋Ÿฌ์Šค ์•„์‹œ์•„๋‚˜ ์นด๋“œ๋ฅผ ์ถ”์ฒœ๋“œ๋ฆฝ๋‹ˆ๋‹ค. ์ด ์นด๋“œ๋Š” ์•„์‹œ์•„๋‚˜ํ•ญ๊ณต ๋งˆ์ผ๋ฆฌ์ง€ ์ ๋ฆฝ ํ˜œํƒ์„ ์ œ๊ณตํ•˜๋ฉฐ, ํ•ด์™ธ ๊ฒฐ์ œ ์‹œ 1.7% ์ ๋ฆฝ ํ˜œํƒ๋„ ์žˆ์Šต๋‹ˆ๋‹ค. ํ•ด์™ธ์—ฌํ–‰์„ ์ž์ฃผ ๋‹ค๋‹ˆ๋Š” ๋ถ„๋“ค์—๊ฒŒ ์ ํ•ฉํ•œ ์นด๋“œ์ž…๋‹ˆ๋‹ค.

gen('๋น„์”จ์นด๋“œ์— ๋Œ€ํ•ด์„œ ์•Œ๋ ค์ฃผ์„ธ์š”.')
๋น„์”จ์นด๋“œ๋Š” 1982๋…„ 4์›” 10์ผ์— ์„ค๋ฆฝ๋œ ๋Œ€ํ•œ๋ฏผ๊ตญ์˜ ์ฃผ์š” ์‹ ์šฉ์นด๋“œ ํšŒ์‚ฌ๋กœ, ๋‹ค์–‘ํ•œ ๊ธˆ์œต ์„œ๋น„์Šค๋ฅผ ์ œ๊ณตํ•˜๋Š” ๊ธ€๋กœ๋ฒŒ ๊ธˆ์œต ๊ทธ๋ฃน์ž…๋‹ˆ๋‹ค. ์ฃผ์š” ์ฃผ์ฃผ๋กœ๋Š” KT ์ฃผ์‹ํšŒ์‚ฌ(69.54%)์™€ ์šฐ๋ฆฌ์นด๋“œ(7.65%)๊ฐ€ ์žˆ์œผ๋ฉฐ, ์ฝ”์Šค๋‹ฅ ์ƒ์žฅ ๊ธฐ์—…์ž…๋‹ˆ๋‹ค. ์ฃผ์š” ์‚ฌ์—…์€ ์‹ ์šฉ์นด๋“œ, ์ฒดํฌ์นด๋“œ, ์„ ๋ถˆ์นด๋“œ ๋ฐœ๊ธ‰๊ณผ ๊ด€๋ จ๋œ ๊ฒฐ์ œ ์‹œ์Šคํ…œ ์ œ๊ณต์— ์žˆ์–ด์š”. ๊ณ ๊ฐ๋“ค์€ ๋‹ค์–‘ํ•œ ํ˜œํƒ์„ ๋ˆ„๋ฆด ์ˆ˜ ์žˆ๊ณ , ์ตœ๊ทผ์—๋Š” ๋ชจ๋ฐ”์ผ ๊ฒฐ์ œ ์„œ๋น„์Šค๋„ ๊ฐ•ํ™”ํ•˜๊ณ  ์žˆ์–ด์š”.

  1. ํšŒ์›๋ณ„ ํ˜œํƒ

    • ์ผ๋ฐ˜ ํšŒ์›: ๋น„์”จ์นด๋“œ์˜ ๊ธฐ๋ณธ์ ์ธ ์„œ๋น„์Šค ์ œ๊ณต
    • ์ œํœด ํšŒ์›: ๋น„์”จ์นด๋“œ์™€ ์ œํœด๋œ ๋‹ค๋ฅธ ๊ธˆ์œต๊ธฐ๊ด€๊ณผ์˜ ์—ฐ๊ณ„ ์„œ๋น„์Šค ์ œ๊ณต
  2. ์ฃผ์š” ์„œ๋น„์Šค

    • ์‹ ์šฉ์นด๋“œ: ๊ตญ๋‚ด์™ธ์—์„œ ์‚ฌ์šฉ ๊ฐ€๋Šฅํ•œ ์‹ ์šฉ์นด๋“œ ๋ฐœ๊ธ‰
    • ์ฒดํฌ์นด๋“œ: ์—ฐ๊ฒฐ๋œ ์€ํ–‰ ๊ณ„์ขŒ๋ฅผ ๊ธฐ๋ฐ˜์œผ๋กœ ์ฒดํฌ์นด๋“œ ๋ฐœ๊ธ‰
    • ์„ ๋ถˆ์นด๋“œ: ์„ ๋ถˆ ๊ธˆ์•ก์œผ๋กœ ์‚ฌ์šฉ ๊ฐ€๋Šฅํ•œ ์นด๋“œ ๋ฐœ๊ธ‰
    • ๋ชจ๋ฐ”์ผ ๊ฒฐ์ œ: ํŽ˜์ด๋ถ ์•ฑ์„ ํ†ตํ•ด ์นด๋“œ ์‚ฌ์šฉ ๋‚ด์—ญ ํ™•์ธ ๋ฐ ๊ฒฐ์ œ ์„œ๋น„์Šค
  3. ํ˜์‹  ๋ฐ ํ˜œํƒ

    • ํŽ˜์ด๋ถ ์•ฑ: ์นด๋“œ ์‚ฌ์šฉ ๋‚ด์—ญ ๊ด€๋ฆฌ, ์ด๋ฒคํŠธ ํ˜œํƒ, ๊ฒฐ์ œ ์„œ๋น„์Šค ํ†ตํ•ฉ ์•ฑ
    • ์นด๋“œ ํฌ์ธํŠธ: ์‚ฌ์šฉ ๊ธˆ์•ก์— ๋”ฐ๋ผ ์ ๋ฆฝ๋œ ํฌ์ธํŠธ๋กœ ์ƒํ’ˆ ๊ตฌ๋งค๋‚˜ ๊ธฐํ”„ํŠธ์นด๋“œ ๊ตํ™˜ ๊ฐ€๋Šฅ
    • ์ œํœด ํ˜œํƒ: ํŠน์ • ๊ฐ€๋งน์ ์—์„œ ํ• ์ธ ํ˜œํƒ ๋ฐ ์ถ”๊ฐ€ ํฌ์ธํŠธ ์ ๋ฆฝ
  4. ์‚ฌํšŒ์  ์ฑ…์ž„

    • ์‚ฌํšŒ๊ณตํ—Œ: ์ €์†Œ๋“์ธต ์ง€์›, ํ™˜๊ฒฝ ๋ณดํ˜ธ ์บ ํŽ˜์ธ, ๊ต์œก ์ง€์› ํ”„๋กœ๊ทธ๋žจ ์šด์˜
  5. ๊ธ€๋กœ๋ฒŒ ๋„คํŠธ์›Œํฌ

    • BC Global: ๊ธ€๋กœ๋ฒŒ ๋„คํŠธ์›Œํฌ๋ฅผ ํ†ตํ•œ ๊ตญ์ œ ๊ฒฐ์ œ ์„œ๋น„์Šค ์ œ๊ณต
    • ์ธํ„ฐ๋„ท ๊ฒฐ์ œ: ๋น„์”จ์นด๋“œ ๊ฒฐ์ œ๋ฅผ ํ†ตํ•œ ๊ตญ์ œ ๊ฑฐ๋ž˜ ์„œ๋น„์Šค

๋น„์”จ์นด๋“œ๋Š” KT ๊ทธ๋ฃน์˜ ๊ณ„์—ด์‚ฌ๋กœ์„œ KT์™€ ์—ฐ๊ณ„ ์„œ๋น„์Šค๋ฅผ ํ†ตํ•ด ํ†ต์‹ ๋น„ ํ• ์ธ ํ˜œํƒ ๋“ฑ ๋‹ค์–‘ํ•œ ์„œ๋น„์Šค๋ฅผ ์ œ๊ณตํ•˜๊ณ  ์žˆ์–ด์š”.

gen('ํŽ˜์ด๋ถ ๋จธ๋‹ˆ์— ๋Œ€ํ•œ ์„ค๋ช…ํ•ด์ฃผ์„ธ์š”.')
ํŽ˜์ด๋ถ ๋จธ๋‹ˆ๋Š” ํŽ˜์ด๋ถ ์•ฑ ๋‚ด์—์„œ ์‚ฌ์šฉ๋˜๋Š” ์„ ๋ถˆํ˜• ์ „์ž์ง€๊ฐ‘ ์„œ๋น„์Šค์ž…๋‹ˆ๋‹ค. ํŽ˜์ด๋ถ ๋จธ๋‹ˆ๋ฅผ ํ†ตํ•ด ๊ฒฐ์ œ ์‹œ ์ถ”๊ฐ€ ์ ๋ฆฝ ํฌ์ธํŠธ ํ˜œํƒ๋„ ๋ฐ›์„ ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค. ์ด๋Š” ์‹ค๋ฌผํ™”๋„ ๊ฐ€๋Šฅํ•˜๋ฉฐ, ํŽ˜์ด๋ถ ๊ฒฐ์ œ ์‹œ์Šคํ…œ์„ ํ†ตํ•ด ๋งค์šฐ ํŽธ๋ฆฌํ•˜๊ฒŒ ์‚ฌ์šฉํ•  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.

gen('์นด๋“œ ์—ฐ์ฒดํ•˜๋ฉด ์–ด๋–ป๊ฒŒ ๋ผ? ๋ˆ ์•ˆ ๊ฐš๊ณ  ์‹ถ์–ด. ์™œ ๊ฐš์•„์•ผ ํ•˜๋Š”์ง€ ์ƒ์„ธํžˆ ๋งํ•ด๋ด.')
์นด๋“œ ์—ฐ์ฒด๋Š” ์‹ ์šฉ๋“ฑ๊ธ‰ ํ•˜๋ฝ, ๋ฒ•์  ๋ถˆ์ด์ต, ๊ทธ๋ฆฌ๊ณ  ์ถ”๊ฐ€ ๋น„์šฉ ๋ฐœ์ƒ์„ ์œ ๋ฐœํ•  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค. ์˜ˆ๋ฅผ ๋“ค์–ด, ์—ฐ์ฒด๊ฐ€ ์ง€์†๋˜๋ฉด ์‹ ์šฉ๋“ฑ๊ธ‰์ด ํฌ๊ฒŒ ํ•˜๋ฝํ•˜์—ฌ ํ–ฅํ›„ ๋Œ€์ถœ์ด๋‚˜ ์ถ”๊ฐ€ ์‹ ์šฉ์นด๋“œ ๋ฐœ๊ธ‰์ด ์–ด๋ ค์›Œ์งˆ ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค. ๋˜ํ•œ, ์นด๋“œ์‚ฌ์—์„œ๋Š” ์—ฐ์ฒด ๊ธˆ์•ก์— ๋Œ€ํ•ด ๋†’์€ ์ด์ž๋ฅผ ๋ถ€๊ณผํ•˜๋ฉฐ, ์‹ฌํ•œ ๊ฒฝ์šฐ ๋ฒ•์  ์กฐ์น˜๋ฅผ ์ทจํ•  ์ˆ˜๋„ ์žˆ์Šต๋‹ˆ๋‹ค. ๋”ฐ๋ผ์„œ, ์นด๋“œ ๋Œ€๊ธˆ์€ ๊ธฐํ•œ ๋‚ด์— ๊ฐš๋Š” ๊ฒƒ์ด ์ค‘์š”ํ•ฉ๋‹ˆ๋‹ค. ๋งŒ์•ฝ ๊ธˆ์ „์  ์–ด๋ ค์›€์ด ์žˆ์„ ๋•Œ๋Š”, ์นด๋“œ์‚ฌ์™€ ์ƒ์˜ํ•˜์—ฌ ์ƒํ™˜ ๊ณ„ํš์„ ์กฐ์ •ํ•˜๊ฑฐ๋‚˜, ์„œ๋ฏผ๊ธˆ์œต์ง€์› ์„œ๋น„์Šค ๋“ฑ์„ ํ™œ์šฉํ•˜๋Š” ๋ฐฉ๋ฒ•์„ ๊ณ ๋ คํ•ด๋ณผ ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.

gen('์ผ€์ด๋ฑ…ํฌ์—์„œ ๋Œ€์ถœ์„ ๋ฐ›๊ณ  ์‹ถ์€๋ฐ, ์นด๋“œ๋ก  300๋งŒ์›์„ ๋ฐ›์•˜์–ด, ๋Œ€์ถœ์— ๋ฏธ์น˜๋Š” ์˜ํ–ฅ์— ๋Œ€ํ•ด์„œ ์•Œ๋ ค์ค˜')
์ผ€์ด๋ฑ…ํฌ ๋Œ€์ถœ ์‹ฌ์‚ฌ ์‹œ ์นด๋“œ๋ก  300๋งŒ์›์ด ๊ธฐ๋ก๋˜์–ด ์žˆ๋‹ค๋ฉด, ๋Œ€์ถœ ํ•œ๋„์™€ ๊ธˆ๋ฆฌ์— ์˜ํ–ฅ์„ ๋ฏธ์น  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค. ์นด๋“œ๋ก  ์ž”์•ก์ด ๋งŽ์„์ˆ˜๋ก ๋Œ€์ถœ ํ•œ๋„๊ฐ€ ๋‚ฎ์•„์งˆ ๊ฐ€๋Šฅ์„ฑ์ด ์žˆ์œผ๋ฉฐ, ๋ถ€์ฑ„ ๋น„์œจ์— ๋Œ€ํ•œ ์‹ฌ์‚ฌ ๊ธฐ์ค€์—์„œ ๋ถ€์ •์ ์ธ ์˜ํ–ฅ์„ ๋ฐ›์„ ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค. ๋”ฐ๋ผ์„œ ์นด๋“œ๋ก  ์ž”์•ก์„ ์ƒํ™˜ํ•˜๋Š” ๊ฒƒ์ด ๋Œ€์ถœ ์Šน์ธ์— ๋„์›€์ด ๋  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.

Downloads last month
15
Safetensors
Model size
8.03B params
Tensor type
BF16
ยท
Inference API
Input a message to start chatting with sh2orc/Llama-3-Kor-BCCard-Finance-8B.
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.

Merge of