distil-gemma-2-2b / README.md
macadeliccc's picture
Update README.md
2b025ae verified
---
base_model: google/gemma-2-2b
datasets: mlabonne/TheTome
---
# Distil Gemma 2 2b
This model is a gemma 2 2b model distilled from google/gemma-2-9b-it and finetuned on the tome.
![image/webp](https://cdn-uploads.huggingface.co/production/uploads/6455cc8d679315e4ef16fbec/89XFihSa8o08wWw8w53uh.webp)
## Prompt Template
ChatML
```
<|im_start|>system
{system}<|im_end|>
<|im_start|>user
{user}<|im_end|>
<|im_start|>assistant
```
## Training Information
This model trained on 8x Nvidia H100 NVL for the equivalent of 120 GPU hours.
+ Loss Achieved: 0.27
+ Epochs: 3
Checkpoints are available in the repo to continue training
## Evals
IN PROGRESS