AlekseiPravdin commited on
Commit
efdabf8
·
verified ·
1 Parent(s): 6a7e874

Upload folder using huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +11 -13
README.md CHANGED
@@ -35,23 +35,21 @@ dtype: float16
35
 
36
  ## Model Features
37
 
38
- This merged model combines the advanced capabilities of the HQQ quantized version of Llama-3.1-8B-Instruct with the fine-tuned prowess of the SauerkrautLM variant. The result is a versatile model that excels in both generative tasks and nuanced understanding of multilingual contexts, particularly in German and English. The model is designed to handle a variety of text generation tasks, making it suitable for applications ranging from conversational agents to content creation.
39
 
40
  ## Evaluation Results
41
 
42
- The individual models have demonstrated impressive performance across various benchmarks:
43
 
44
- - **mobiuslabsgmbh/Llama-3.1-8b-instruct_4bitgs64_hqq_calib**:
45
- - ARC (25-shot): 60.49
46
- - HellaSwag (10-shot): 80.16
47
- - MMLU (5-shot): 68.98
48
- - Average performance: 69.51
49
-
50
- - **VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct**:
51
- - Fine-tuned on German-English data, showcasing significant improvements in multilingual capabilities.
52
-
53
- The merged model inherits the strengths of both parent models, providing enhanced performance in multilingual contexts while maintaining the efficiency of the HQQ quantization.
54
 
55
  ## Limitations
56
 
57
- While the merged model benefits from the strengths of both parent models, it may also carry over some limitations. For instance, the potential for uncensored content remains a concern, as noted in the SauerkrautLM documentation. Additionally, the model's performance may vary depending on the specific task and the quality of the input data. Users should be aware of these factors when deploying the model in real-world applications.
 
 
 
35
 
36
  ## Model Features
37
 
38
+ This merged model combines the advanced capabilities of the HQQ quantized version of Llama-3.1-8B-Instruct with the fine-tuned prowess of Llama-3.1-SauerkrautLM-8b-Instruct. The result is a versatile model that excels in both generative tasks and nuanced understanding of multilingual contexts, particularly in German and English. The integration of Spectrum Fine-Tuning from the SauerkrautLM model enhances its efficiency and performance, making it suitable for a wide range of applications.
39
 
40
  ## Evaluation Results
41
 
42
+ The evaluation results from the parent models indicate strong performance across various benchmarks. For instance, the HQQ 4-bit version of Llama-3.1-8B-Instruct achieved notable scores on tasks like ARC, HellaSwag, and MMLU, while the SauerkrautLM model demonstrated significant improvements in multilingual capabilities. The combined strengths of these models in the merged version are expected to yield even better results in similar tasks.
43
 
44
+ | Benchmark | Llama-3.1-8B-Instruct (HQQ 4-bit) | Llama-3.1-SauerkrautLM |
45
+ |-------------------|-----------------------------------|-------------------------|
46
+ | ARC (25-shot) | 60.32 | Not specified |
47
+ | HellaSwag (10-shot)| 79.21 | Not specified |
48
+ | MMLU (5-shot) | 67.07 | Not specified |
49
+ | Average | 68.00 | Not specified |
 
 
 
 
50
 
51
  ## Limitations
52
 
53
+ While the merged model benefits from the strengths of both parent models, it may also inherit some limitations. For instance, the HQQ quantization process could introduce certain biases or inaccuracies in specific contexts, and the fine-tuning on a limited dataset may not cover all nuances of the languages involved. Users should be aware of these potential issues and exercise caution when deploying the model in sensitive applications. Additionally, despite efforts to ensure appropriate behavior, the possibility of encountering uncensored content remains.
54
+
55
+ In summary, Llama-3.1-8b-instruct_4bitgs64_hqq_calib-Llama-3.1-SauerkrautLM-8b-Instruct-dare-merge represents a significant advancement in language modeling, combining the best features of its predecessors while also carrying forward some of their limitations.