File size: 1,078 Bytes
6438b36 09245b4 6438b36 09245b4 6438b36 60cc967 6438b36 60cc967 6438b36 60cc967 6438b36 60cc967 6438b36 d10e001 6438b36 09245b4 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 |
---
library_name: transformers
license: apache-2.0
language:
- ko
---
# Model Card for Model ID
## Model Details
### Model Description
The Gemma Self-Attention Merged model is a large language model created by merging the self-attention layers of an [English-based Gemma 7B model](https://huggingface.co/google/gemma-1.1-7b-it) and a [Korean-based Gemma 7B model](beomi/gemma-ko-7b). This merger allows the model to leverage the capabilities of both the English and Korean models, resulting in a more versatile and capable language model that can perform well on tasks involving both English and Korean text.
The key features of this merged model include:
- Increased self-attention capacity with doubled number of attention heads
- Ability to handle both English and Korean language input
- Potential for improved performance on a wide range of natural language processing tasks
#### Chat template
**system:** system message...
**B:** user message...
**A:** assistant message...
### Model Sources
- **Repository:** https://github.com/lcw99/merge-gemma-attn.git |