Update Log

  • 2023.12.14: Initial Release of Open-Llama-2-Ko

Open-Llama-2-Ko πŸ¦™πŸ‡°πŸ‡·

Open-Llama-2-Ko represents an advanced iteration of the Llama 2 model, featuring an expanded vocabulary and the inclusion of a Korean corpus for enhanced pretraining. Similar to its predecessor, Llama-2-Ko, this model operates within the range of generative text models, with parameter counts ranging from 7 billion to 70 billion. The focus of this repository is on the 7B pretrained version, designed to integrate seamlessly with the Hugging Face Transformers format.

The primary distinction between the Llama-2-Ko Series and Open-Llama-2-Ko lies in the dataset. Open-Llama-2-Ko exclusively utilizes publicly accessible Korean corpora, including sources such as AI Hub, Modu Corpus, λͺ¨λ‘μ˜ λ§λ­‰μΉ˜, and Korean Wikipedia.

As training was conducted solely with publicly available corpora, this model is open for unrestricted use by everyone, adhering to the MIT License*.

*MIT License under LLAMA 2 COMMUNITY LICENSE AGREEMENT

Model Details

Model Developers: Junbum Lee (Beomi)

Variations: Open-Llama-2-Ko will be available in different parameter sizes β€” 7B and 13B β€” along with various pretrained options.

Input: The model accepts only text input.

Output: The model produces text output exclusively.

Model Architecture:

Open-Llama-2-Ko is an auto-regressive language model that leverages an optimized transformer architecture derived from Llama-2.

Training Data Parameters Content Length GQA Tokens Learning Rate
Llama 2 A curated mix of Publicly Accessible Korean Corpora 7B 2k ✘ >15B* 5e-5

Training Corpus

The model was trained using selected datasets from AIHub and Modu Corpus. Detailed information about the training datasets is available below:

  • AI Hub: corpus/AI_HUB
    • Only the Training segment of the data was used.
    • The Validation and Test segments were deliberately excluded.
  • Modu Corpus: corpus/MODU_CORPUS

The final JSONL dataset used to train this model is approximately 61GB in size.

Total token count: Approximately 15 billion tokens (*using the expanded tokenizer. With the original Llama tokenizer, >60 billion tokens.)

Vocab Expansion

Model Name Vocabulary Size Description
Original Llama-2 32000 Sentencepiece BPE
Expanded Llama-2-Ko 46336 Sentencepiece BPE. Added Korean vocab and merges

Tokenizing "μ•ˆλ…•ν•˜μ„Έμš”, μ˜€λŠ˜μ€ 날씨가 μ’‹λ„€μš”."

Model Tokens
Llama-2 ['▁', 'μ•ˆ', '<0xEB>', '<0x85>', '<0x95>', 'ν•˜', 'μ„Έ', 'μš”', ',', '▁', '였', '<0xEB>', '<0x8A>', '<0x98>', '은', '▁', '<0xEB>', '<0x82>', '<0xA0>', '씨', 'κ°€', '▁', '<0xEC>', '<0xA2>', '<0x8B>', '<0xEB>', '<0x84>', '<0xA4>', 'μš”']
Llama-2-Ko ['β–μ•ˆλ…•', 'ν•˜μ„Έμš”', ',', 'β–μ˜€λŠ˜μ€', '▁날', '씨가', 'β–μ’‹λ„€μš”']

Tokenizing "Llama 2: Open Foundation and Fine-Tuned Chat Models"

Model Tokens
Llama-2 ['▁L', 'l', 'ama', '▁', '2', ':', '▁Open', '▁Foundation', '▁and', '▁Fine', '-', 'T', 'un', 'ed', '▁Ch', 'at', '▁Mod', 'els']
Llama-2-Ko ['▁L', 'l', 'ama', '▁', '2', ':', '▁Open', '▁Foundation', '▁and', '▁Fine', '-', 'T', 'un', 'ed', '▁Ch', 'at', '▁Mod', 'els']

LICENSE

MIT License under LLAMA 2 COMMUNITY LICENSE AGREEMENT

Model Benchmark

LM Eval Harness - Korean (polyglot branch)

TBD

Citation

TBD

Acknowledgements

Downloads last month
2,817
Safetensors
Model size
6.86B params
Tensor type
BF16
Β·
Inference Examples
Inference API (serverless) has been turned off for this model.

Model tree for beomi/open-llama-2-ko-7b

Adapters
8 models
Finetunes
3 models
Quantizations
1 model

Collection including beomi/open-llama-2-ko-7b