--- license: apache-2.0 language: - en - zh --- license: apache-2.0 --- # GreenBit Yi This is GreenBitAI's pretrained **4-bit** Yi 34B model with extreme compression yet still strong performance. Please refer to our [Github page](https://github.com/GreenBitAI/low_bit_llama/tree/low_bit_yi) for the code to run the model and more information. ## Model Description - **Developed by:** [GreenBitAI](https://github.com/GreenBitAI) - **Evaluated By:** 01-Yi official - **Model type:** Causal (Llama 2/Yi 6B) - **Language(s) (NLP):** English - **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0), [Llama 2 license agreement](https://ai.meta.com/resources/models-and-libraries/llama-downloads/) ## Few Shot Evaluation (officially evaluated by 01-Yi) | Model | Yi-34B FP16| [Yi-34B 4 bit](https://huggingface.co/GreenBitAI/yi-34b-w4a16g32) | Yi-6B FP16 | [Yi-6B 4 bit](https://huggingface.co/GreenBitAI/yi-6b-w4a16g32) | |----------------|-----------|----------|----------|---------| | GroupSize | - | 32 | - | 8 | | Model Size (GB)| 68.79 | 19.89 | 12.12 | 4.04 | | AVG | 70.64 | 69.7 | 60.11 | 59.14 | | **Detailed Evaluation** | | | | | | MMLU | 76.32 | 75.42 | 63.24 | 62.09 | | CMMLU | 83.65 | 83.07 | 75.53 | 72.85 | | ARC-e | 84.42 | 84.13 | 77.23 | 76.52 | | ARC-c | 61.77 | 59.56 | 50.34 | 48.47 | | GAOKAO | 82.8 | 81.37 | 72.2 | 72.87 | | GSM8K | 67.24 | 63.61 | 32.52 | 28.05 | | HumanEval | 25.6 | 25 | 15.85 | 15.85 | | BBH | 54.3 | 52.3 | 42.8 | 41.47 | | WinoGrande | 78.68 | 78.53 | 70.63 | 71.19 | | PIQA | 82.86 | 82.75 | 78.56 | 79.05 | | SIQA | 74.46 | 73.44 | 64.53 | 64.53 | | HellaSwag | 83.64 | 83.02 | 74.91 | 73.27 | | OBQA | 91.6 | 90.8 | 85.4 | 82.6 | | CSQA | 83.37 | 83.05 | 76.9 | 75.43 | | TriviaQA | 81.52 | 80.73 | 64.85 | 61.75 | | SquAD | 92.46 | 91.12 | 88.95 | 88.39 | | BoolQ | 88.25 | 88.17 | 76.23 | 77.1 | | MBPP | 41 | 39.68 | 26.32 | 25.13 | | QUAC | 48.61 | 47.43 | 40.92 | 40.16 | | Lambda | 73.18 | 73.39 | 67.74 | 67.8 | | NaturalQuestion| 27.67 | 27.21 | 16.69 | 17.42 | # Zero Shot Evaluation | Task | Metric | Yi-6B FP16 | [Yi-6B 4 bit](https://huggingface.co/GreenBitAI/yi-6b-w4a16g32) | [Yi-34B 4 bit](https://huggingface.co/GreenBitAI/yi-34b-w4a16g32) | |---------------|--------|---------|-------------|--------------| | Openbookqa | acc | 0.314 | 0.324 | 0.344 | | | ac_norm| 0.408 | 0.42 | 0.474 | | arc_challenge | acc | 0.462 | 0.4573 | 0.569 | | | ac_norm| 0.504 | 0.483 | 0.5964 | | hellawswag | acc | 0.553 | 0.5447 | 0.628 | | | ac_norm| 0.749 | 0.7327 | 0.83 | | piqa | acc | 0.777 | 0.7709 | 0.8079 | | | ac_norm| 0.787 | 0.7894 | 0.828 | | arc_easy | acc | 0.777 | 0.7697 | 0.835 | | | ac_norm| 0.774 | 0.7659 | 0.84 | | Winogrande | acc | 0.707 | 0.7095 | 0.7853 | | boolq | acc | 0.755 | 0.7648 | 0.886 | | truthfulqa_mc | mc1 | 0.29 | 0.2729 | 0.4026 | | | mc2 | 0.419 | 0.4033 | 0.5528 | | anli_r1 | acc | 0.423 | 0.416 | 0.554 | | anli_r2 | acc | 0.409 | 0.409 | 0.518 | | anli_r3 | acc | 0.411 | 0.393 | 0.4983 | | wic | acc | 0.529 | 0.545 | 0.5376 | | rte | acc | 0.685 | 0.7039 | 0.7617 | | record | f1 | 0.904 | 0.9011 | 0.924 | | | em | 0.8962 | 0.8927 | 0.916 | | Average | | 0.596 | 0.5937 | 0.6708 |