--- license: other license_name: tongyi-qianwen license_link: https://huggingface.co/Qwen/Qwen2-72B-Instruct/resolve/main/LICENSE --- This model is an improved version for Korean, based on the [Qwen2-72B-Instruct](https://huggingface.co/Qwen/Qwen2-72B-Instruct) model. ### LogicKor Benchmark (24.07.31) * [The following benchmark](https://lk.instruct.kr/) ranks are based on 1-shot evaluation. | Rank | Model | Reasoning | Math | Writing | Coding | Understanding | Grammar | Singleturn | Multiturn | Total | Parameters | |------|-------|-----------|-------|--------|--------|-------|---------|-----------|-----------|-------|---------| | 1 | openai/gpt-4o-2024-05-13 | 9.21 | 8.71 | 9.64 | 9.78 | 9.64 | 9.50 | 9.33 | 9.50 | 9.41 | ? | | 2 | anthropic/claude-3-5-sonnet-20240620 | 8.64 | 8.42 | 9.85 | 9.78 | 9.92 | 9.21 | 9.26 | 9.35 | 9.30 | ? | | 4 | mistralai/Mistral-Large-Instruct-2407 | 9.71 | 9.07 | 9.57 | 9.92 | 9.92 | 6.78 | 9.19 | 9.14 | 9.16 | 123B | | 8 | meta-llama/Meta-Llama-3.1-405B-Instruct-FP8 | 8.78 | 7.14 | 9.28 | 9.64 | 9.64| 8.57 | 8.97 | 8.71 | 8.84 | 405B | | 9 | ```denial07/Qwen2-72B-Instruct-kor-dpo``` | 8.85 | 8.21 | 9.14 | 9.71 | 9.64 | 7.21 | 8.88 | 8.71 | 8.79 | 72B | | 10 | Qwen/Qwen2-72B-Instruct | 8.00 | 8.14 | 9.07 | 9.85 | 9.78 | 7.28 | 8.61 | 8.76 | 8.69 | 72B | | 11 | google/gemini-1.5-pro-001 | 7.00 | 8.00 | 9.57 | 8.85 | 9.35 | 8.64 | 8.61 | 8.52 | 8.57 | ? | ### KMMLU Benchmark * [HAERAE-HUB/KMMLU](https://huggingface.co/datasets/HAERAE-HUB/KMMLU) benchmark accuracy score. | Category |Qwen2-72B-it-kor-dpo| Qwen2-72B-it | Mistral-Large-it-2407 | Questions | |-----------------|--------------------|---------------|-----------------------|------------| | HUMSS | 0.63 | 0.63 | 0.62 | 5130 | | STEM | 0.59 | 0.59 | 0.57 | 9900 | | Applied Science | 0.56 | 0.56 | 0.54 | 11600 | | Other | 0.58 | 0.58 | 0.54 | 8400 | | Overall Accuracy| 0.58 | 0.58 | 0.56 | 35030 |