|
--- |
|
license: apache-2.0 |
|
language: |
|
- en |
|
--- |
|
<div align="center"> |
|
|
|
# KobbleTiny-1.1B |
|
</div> |
|
|
|
This is a finetune of https://huggingface.co/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T trained on a small 50mb subset of the Kobble Dataset. |
|
Training was done in under 2 hours on a single Nvidia RTX 2060 Mobile GPU with qLora (LR 1.5e-4, rank 8, alpha 16, batch size 1, gradient acc. 4, 1024 ctx). |
|
|
|
## Dataset and Objectives |
|
|
|
The Kobble Dataset is a semi-private aggregated dataset made from multiple online sources and web scrapes. |
|
It contains content chosen and formatted specifically to work with KoboldAI software and Kobold Lite. |
|
|
|
#### Dataset Categories: |
|
- Instruct: Single turn instruct examples presented in the Alpaca format, with an emphasis on uncensored and unrestricted responses. |
|
- Chat: Two participant roleplay conversation logs in a multi-turn raw chat format that KoboldAI uses. |
|
- Story: Unstructured fiction excerpts, including literature containing various erotic and provocative content. |
|
|
|
<!-- prompt-template start --> |
|
## Prompt template: Alpaca |
|
|
|
``` |
|
### Instruction: |
|
{prompt} |
|
|
|
### Response: |
|
``` |
|
|
|
<!-- prompt-template end --> |
|
|
|
**Note:** *No assurances will be provided about the **origins, safety, or copyright status** of this model, or of **any content** within the Kobble dataset.* |
|
*If you belong to a country or organization that has strict AI laws or restrictions against unlabelled or unrestricted content, you are advised not to use this model.* |
|
|