--- license: llama3 --- # Alpesteibock-Llama-3-8B-Alpha **Alpesteibock-Llama-3-8B-Alpha** is an experimental QLoRA fine-tune of [NousResearch/Hermes-2-Pro-Llama-3-8B](https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-8B) on a dataset of more than 28 million tokens of Swiss German text from multiple sources. ## Dataset ## Training Details Hardware: 1x RTX 4090 Duration: ~30 hours in total (~2 hours for first phase and ~28 hours for second phase) ### Hyperparameters Adapter: QLoRA Precision: 4 bit Optimizer: adamw_bnb_8bit LoRA Rank: 256 LoRA Alpha: 256 Learning Rate: 1e-5 Context Length: 4096 tokens Batch Size: 1 Gradient Accumulation Steps: 1 Sample Packing: Off for first phase, on for second phase Epochs: 2 ## Limitations