Mistral-7B-0.3_auto
Collection
36 items
•
Updated
This model is a fine-tuned version of mistralai/Mistral-7B-Instruct-v0.3 on the GaetanMichelet/chat-60_ft_task-3_auto dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
1.2067 | 0.8696 | 5 | 1.1800 |
1.1313 | 1.9130 | 11 | 1.0719 |
1.0437 | 2.9565 | 17 | 0.9889 |
0.8887 | 4.0 | 23 | 0.8997 |
0.8094 | 4.8696 | 28 | 0.8760 |
0.7792 | 5.9130 | 34 | 0.8672 |
0.7139 | 6.9565 | 40 | 0.8654 |
0.666 | 8.0 | 46 | 0.8719 |
0.6045 | 8.8696 | 51 | 0.8943 |
0.5506 | 9.9130 | 57 | 0.9271 |
0.4489 | 10.9565 | 63 | 0.9576 |
0.3597 | 12.0 | 69 | 1.0070 |
0.3326 | 12.8696 | 74 | 1.0801 |
0.2557 | 13.9130 | 80 | 1.2003 |
Base model
mistralai/Mistral-7B-v0.3