Update README.md
Browse files
README.md
CHANGED
@@ -47,8 +47,9 @@ The model **SauerkrautLM-Qwen-32b** is a **joint effort** between **VAGO solutio
|
|
47 |
- **Contact:** [VAGO solutions](https://vago-solutions.ai), [Hyperspace.ai](https://hyperspace.computer/)
|
48 |
|
49 |
### Training procedure:
|
50 |
-
We trained this model for 2 epochs on 160k data samples with SFT.
|
51 |
-
Afterwards we applied DPO for 1 epoch with 110k data.
|
|
|
52 |
|
53 |
**We teached German language skills on this model.** As far as we know, it is the first Qwen 32B model with bilingual skills in German and English. Nevertheless, formulations may occur that are not entirely correct (still work in progress).
|
54 |
|
|
|
47 |
- **Contact:** [VAGO solutions](https://vago-solutions.ai), [Hyperspace.ai](https://hyperspace.computer/)
|
48 |
|
49 |
### Training procedure:
|
50 |
+
- We trained this model for 2 epochs on 160k data samples with SFT.
|
51 |
+
- Afterwards we applied DPO for 1 epoch with 110k data.
|
52 |
+
- LaserRMT version coming soon
|
53 |
|
54 |
**We teached German language skills on this model.** As far as we know, it is the first Qwen 32B model with bilingual skills in German and English. Nevertheless, formulations may occur that are not entirely correct (still work in progress).
|
55 |
|