--- license: apache-2.0 language: - en ---
# KobbleTinyV2-1.1B
This is a finetune of https://huggingface.co/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T trained on a small 50mb subset of the Kobble Dataset. Training was done in under 2 hours on a single Nvidia RTX 2060 Mobile GPU with qLora (LR 1.5e-4, rank 8, alpha 16, batch size 2, gradient acc. 4, 2048 ctx). You can obtain the GGUF quantization of this model here: https://huggingface.co/concedo/KobbleTinyV2-1.1B-GGUF Update: KobbleTiny has been upgraded to V2! The old V1 is [still available at this link](https://huggingface.co/concedo/KobbleTiny/tree/eb0c96864bfecfd6ac9ece1a42c4654b4997eb72). Try it live now: https://concedo-koboldcpp-kobbletiny.hf.space/ ## Dataset and Objectives The Kobble Dataset is a semi-private aggregated dataset made from multiple online sources and web scrapes. It contains content chosen and formatted specifically to work with KoboldAI software and Kobold Lite. #### Dataset Categories: - Instruct: Single turn instruct examples presented in the Alpaca format, with an emphasis on uncensored and unrestricted responses. - Chat: Two participant roleplay conversation logs in a multi-turn raw chat format that KoboldAI uses. - Story: Unstructured fiction excerpts, including literature containing various erotic and provocative content. ## Prompt template: Alpaca ``` ### Instruction: {prompt} ### Response: ``` **Note:** *No assurances will be provided about the **origins, safety, or copyright status** of this model, or of **any content** within the Kobble dataset.* *If you belong to a country or organization that has strict AI laws or restrictions against unlabelled or unrestricted content, you are advised not to use this model.*